Started by upstream project "integration-distribution-test-calcium" build number 227 originally caused by: Started by upstream project "autorelease-release-scandium-mvn38-openjdk17" build number 12 originally caused by: Started by timer Running as SYSTEM [EnvInject] - Loading node environment variables. Building remotely on prd-centos8-robot-2c-8g-10017 (centos8-robot-2c-8g) in workspace /w/workspace/openflowplugin-csit-3node-clustering-only-calcium [ssh-agent] Looking for ssh-agent implementation... [ssh-agent] Exec ssh-agent (binary ssh-agent on a remote machine) $ ssh-agent SSH_AUTH_SOCK=/tmp/ssh-8EinDYD3VjtH/agent.5284 SSH_AGENT_PID=5285 [ssh-agent] Started. Running ssh-add (command line suppressed) Identity added: /w/workspace/openflowplugin-csit-3node-clustering-only-calcium@tmp/private_key_15056833576193002484.key (/w/workspace/openflowplugin-csit-3node-clustering-only-calcium@tmp/private_key_15056833576193002484.key) [ssh-agent] Using credentials jenkins (Release Engineering Jenkins Key) The recommended git tool is: NONE using credential opendaylight-jenkins-ssh Wiping out workspace first. Cloning the remote Git repository Cloning repository git://devvexx.opendaylight.org/mirror/integration/test > git init /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test # timeout=10 Fetching upstream changes from git://devvexx.opendaylight.org/mirror/integration/test > git --version # timeout=10 > git --version # 'git version 2.43.0' using GIT_SSH to set credentials Release Engineering Jenkins Key [INFO] Currently running in a labeled security context [INFO] Currently SELinux is 'enforcing' on the host > /usr/bin/chcon --type=ssh_home_t /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test@tmp/jenkins-gitclient-ssh10682333066772749822.key Verifying host key using known hosts file You're using 'Known hosts file' strategy to verify ssh host keys, but your known_hosts file does not exist, please go to 'Manage Jenkins' -> 'Security' -> 'Git Host Key Verification Configuration' and configure host key verification. > git fetch --tags --force --progress -- git://devvexx.opendaylight.org/mirror/integration/test +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url git://devvexx.opendaylight.org/mirror/integration/test # timeout=10 > git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10 > git config remote.origin.url git://devvexx.opendaylight.org/mirror/integration/test # timeout=10 Fetching upstream changes from git://devvexx.opendaylight.org/mirror/integration/test using GIT_SSH to set credentials Release Engineering Jenkins Key [INFO] Currently running in a labeled security context [INFO] Currently SELinux is 'enforcing' on the host > /usr/bin/chcon --type=ssh_home_t /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test@tmp/jenkins-gitclient-ssh15953570683248604651.key Verifying host key using known hosts file You're using 'Known hosts file' strategy to verify ssh host keys, but your known_hosts file does not exist, please go to 'Manage Jenkins' -> 'Security' -> 'Git Host Key Verification Configuration' and configure host key verification. > git fetch --tags --force --progress -- git://devvexx.opendaylight.org/mirror/integration/test master # timeout=10 > git rev-parse FETCH_HEAD^{commit} # timeout=10 Checking out Revision 02dd570f7544c8cd6eb17bda775de9bccc89f923 (origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f 02dd570f7544c8cd6eb17bda775de9bccc89f923 # timeout=10 Commit message: "Migrate Get Requests invocations(libraries)" > git rev-parse FETCH_HEAD^{commit} # timeout=10 > git rev-list --no-walk c88ead2b4a65ba004398ad07cfe033ae1e467b17 # timeout=10 No emails were triggered. provisioning config files... copy managed file [npmrc] to file:/home/jenkins/.npmrc copy managed file [pipconf] to file:/home/jenkins/.config/pip/pip.conf copy managed file [clouds-yaml] to file:/home/jenkins/.config/openstack/clouds.yaml [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins7919638128880886349.sh ---> python-tools-install.sh Setup pyenv: system * 3.8.13 (set by /opt/pyenv/version) * 3.9.13 (set by /opt/pyenv/version) * 3.10.13 (set by /opt/pyenv/version) * 3.11.7 (set by /opt/pyenv/version) lf-activate-venv(): INFO: Creating python3 venv at /tmp/venv-z35k lf-activate-venv(): INFO: Save venv in file: /tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH Generating Requirements File Python 3.11.7 pip 24.0 from /tmp/venv-z35k/lib/python3.11/site-packages/pip (python 3.11) appdirs==1.4.4 argcomplete==3.3.0 aspy.yaml==1.3.0 attrs==23.2.0 autopage==0.5.2 beautifulsoup4==4.12.3 boto3==1.34.88 botocore==1.34.88 bs4==0.0.2 cachetools==5.3.3 certifi==2024.2.2 cffi==1.16.0 cfgv==3.4.0 chardet==5.2.0 charset-normalizer==3.3.2 click==8.1.7 cliff==4.6.0 cmd2==2.4.3 cryptography==3.3.2 debtcollector==3.0.0 decorator==5.1.1 defusedxml==0.7.1 Deprecated==1.2.14 distlib==0.3.8 dnspython==2.6.1 docker==4.2.2 dogpile.cache==1.3.2 email_validator==2.1.1 filelock==3.13.4 future==1.0.0 gitdb==4.0.11 GitPython==3.1.43 google-auth==2.29.0 httplib2==0.22.0 identify==2.5.36 idna==3.7 importlib-resources==1.5.0 iso8601==2.1.0 Jinja2==3.1.3 jmespath==1.0.1 jsonpatch==1.33 jsonpointer==2.4 jsonschema==4.21.1 jsonschema-specifications==2023.12.1 keystoneauth1==5.6.0 kubernetes==29.0.0 lftools==0.37.10 lxml==5.2.1 MarkupSafe==2.1.5 msgpack==1.0.8 multi_key_dict==2.0.3 munch==4.0.0 netaddr==1.2.1 netifaces==0.11.0 niet==1.4.2 nodeenv==1.8.0 oauth2client==4.1.3 oauthlib==3.2.2 openstacksdk==3.1.0 os-client-config==2.1.0 os-service-types==1.7.0 osc-lib==3.0.1 oslo.config==9.4.0 oslo.context==5.5.0 oslo.i18n==6.3.0 oslo.log==5.5.1 oslo.serialization==5.4.0 oslo.utils==7.1.0 packaging==24.0 pbr==6.0.0 platformdirs==4.2.0 prettytable==3.10.0 pyasn1==0.6.0 pyasn1_modules==0.4.0 pycparser==2.22 pygerrit2==2.0.15 PyGithub==2.3.0 pyinotify==0.9.6 PyJWT==2.8.0 PyNaCl==1.5.0 pyparsing==2.4.7 pyperclip==1.8.2 pyrsistent==0.20.0 python-cinderclient==9.5.0 python-dateutil==2.9.0.post0 python-heatclient==3.5.0 python-jenkins==1.8.2 python-keystoneclient==5.4.0 python-magnumclient==4.4.0 python-novaclient==18.6.0 python-openstackclient==6.6.0 python-swiftclient==4.5.0 PyYAML==6.0.1 referencing==0.34.0 requests==2.31.0 requests-oauthlib==2.0.0 requestsexceptions==1.4.0 rfc3986==2.0.0 rpds-py==0.18.0 rsa==4.9 ruamel.yaml==0.18.6 ruamel.yaml.clib==0.2.8 s3transfer==0.10.1 simplejson==3.19.2 six==1.16.0 smmap==5.0.1 soupsieve==2.5 stevedore==5.2.0 tabulate==0.9.0 toml==0.10.2 tomlkit==0.12.4 tqdm==4.66.2 typing_extensions==4.11.0 tzdata==2024.1 urllib3==1.26.18 virtualenv==20.25.3 wcwidth==0.2.13 websocket-client==1.7.0 wrapt==1.16.0 xdg==6.0.0 xmltodict==0.13.0 yq==3.4.1 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content OS_STACK_TEMPLATE=csit-2-instance-type.yaml OS_CLOUD=vex OS_STACK_NAME=releng-openflowplugin-csit-3node-clustering-only-calcium-227 OS_STACK_TEMPLATE_DIR=openstack-hot [EnvInject] - Variables injected successfully. provisioning config files... copy managed file [clouds-yaml] to file:/home/jenkins/.config/openstack/clouds.yaml [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins6044806444537325326.sh ---> Create parameters file for OpenStack HOT OpenStack Heat parameters generated ----------------------------------- parameters: vm_0_count: '3' vm_0_flavor: 'v3-standard-4' vm_0_image: 'ZZCI - CentOS Stream 8 - builder - x86_64 - 20240401-160147.446' vm_1_count: '1' vm_1_flavor: 'v3-standard-2' vm_1_image: 'ZZCI - Ubuntu 18.04 - mininet-ovs-28 - x86_64 - 20230601-180106.003' job_name: '57946-227' silo: 'releng' [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash -l /tmp/jenkins9124789466731001677.sh ---> Create HEAT stack + source /home/jenkins/lf-env.sh + lf-activate-venv --python python3 'lftools[openstack]' kubernetes niet python-heatclient python-openstackclient python-magnumclient yq ++ mktemp -d /tmp/venv-XXXX + lf_venv=/tmp/venv-m8Vk + local venv_file=/tmp/.os_lf_venv + local python=python3 + local options + local set_path=true + local install_args= ++ getopt -o np:v: -l no-path,system-site-packages,python:,venv-file: -n lf-activate-venv -- --python python3 'lftools[openstack]' kubernetes niet python-heatclient python-openstackclient python-magnumclient yq + options=' --python '\''python3'\'' -- '\''lftools[openstack]'\'' '\''kubernetes'\'' '\''niet'\'' '\''python-heatclient'\'' '\''python-openstackclient'\'' '\''python-magnumclient'\'' '\''yq'\''' + eval set -- ' --python '\''python3'\'' -- '\''lftools[openstack]'\'' '\''kubernetes'\'' '\''niet'\'' '\''python-heatclient'\'' '\''python-openstackclient'\'' '\''python-magnumclient'\'' '\''yq'\''' ++ set -- --python python3 -- 'lftools[openstack]' kubernetes niet python-heatclient python-openstackclient python-magnumclient yq + true + case $1 in + python=python3 + shift 2 + true + case $1 in + shift + break + case $python in + local pkg_list= + [[ -d /opt/pyenv ]] + echo 'Setup pyenv:' Setup pyenv: + export PYENV_ROOT=/opt/pyenv + PYENV_ROOT=/opt/pyenv + export PATH=/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin + PATH=/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin + pyenv versions system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) + command -v pyenv ++ pyenv init - --no-rehash + eval 'PATH="$(bash --norc -ec '\''IFS=:; paths=($PATH); for i in ${!paths[@]}; do if [[ ${paths[i]} == "'\'''\''/opt/pyenv/shims'\'''\''" ]]; then unset '\''\'\'''\''paths[i]'\''\'\'''\''; fi; done; echo "${paths[*]}"'\'')" export PATH="/opt/pyenv/shims:${PATH}" export PYENV_SHELL=bash source '\''/opt/pyenv/libexec/../completions/pyenv.bash'\'' pyenv() { local command command="${1:-}" if [ "$#" -gt 0 ]; then shift fi case "$command" in rehash|shell) eval "$(pyenv "sh-$command" "$@")" ;; *) command pyenv "$command" "$@" ;; esac }' +++ bash --norc -ec 'IFS=:; paths=($PATH); for i in ${!paths[@]}; do if [[ ${paths[i]} == "/opt/pyenv/shims" ]]; then unset '\''paths[i]'\''; fi; done; echo "${paths[*]}"' ++ PATH=/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin ++ export PATH=/opt/pyenv/shims:/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin ++ PATH=/opt/pyenv/shims:/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin ++ export PYENV_SHELL=bash ++ PYENV_SHELL=bash ++ source /opt/pyenv/libexec/../completions/pyenv.bash +++ complete -F _pyenv pyenv ++ lf-pyver python3 ++ local py_version_xy=python3 ++ local py_version_xyz= ++ grep -E '^[0-9.]*[0-9]$' ++ sed 's/^[ *]* //' ++ awk '{ print $1 }' ++ pyenv versions ++ local command ++ command=versions ++ '[' 1 -gt 0 ']' ++ shift ++ case "$command" in ++ command pyenv versions ++ pyenv versions ++ [[ ! -s /tmp/.pyenv_versions ]] +++ sort -V +++ grep '^3' /tmp/.pyenv_versions +++ tail -n 1 ++ py_version_xyz=3.11.7 ++ [[ -z 3.11.7 ]] ++ echo 3.11.7 ++ return 0 + pyenv local 3.11.7 + local command + command=local + '[' 2 -gt 0 ']' + shift + case "$command" in + command pyenv local 3.11.7 + pyenv local 3.11.7 + for arg in "$@" + case $arg in + pkg_list+='lftools[openstack] ' + for arg in "$@" + case $arg in + pkg_list+='kubernetes ' + for arg in "$@" + case $arg in + pkg_list+='niet ' + for arg in "$@" + case $arg in + pkg_list+='python-heatclient ' + for arg in "$@" + case $arg in + pkg_list+='python-openstackclient ' + for arg in "$@" + case $arg in + pkg_list+='python-magnumclient ' + for arg in "$@" + case $arg in + pkg_list+='yq ' + [[ -f /tmp/.os_lf_venv ]] ++ cat /tmp/.os_lf_venv + lf_venv=/tmp/venv-z35k + echo 'lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from' file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv + /tmp/venv-z35k/bin/python3 -m pip install --upgrade --quiet pip virtualenv + [[ -z lftools[openstack] kubernetes niet python-heatclient python-openstackclient python-magnumclient yq ]] + echo 'lf-activate-venv(): INFO: Installing: lftools[openstack] kubernetes niet python-heatclient python-openstackclient python-magnumclient yq ' lf-activate-venv(): INFO: Installing: lftools[openstack] kubernetes niet python-heatclient python-openstackclient python-magnumclient yq + /tmp/venv-z35k/bin/python3 -m pip install --upgrade --quiet --upgrade-strategy eager 'lftools[openstack]' kubernetes niet python-heatclient python-openstackclient python-magnumclient yq + type python3 + true + echo 'lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH' lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH + PATH=/tmp/venv-z35k/bin:/opt/pyenv/shims:/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin + return 0 + openstack --os-cloud vex limits show --absolute +--------------------------+---------+ | Name | Value | +--------------------------+---------+ | maxTotalInstances | -1 | | maxTotalCores | 450 | | maxTotalRAMSize | 1000000 | | maxServerMeta | 128 | | maxImageMeta | 128 | | maxPersonality | 5 | | maxPersonalitySize | 10240 | | maxTotalKeypairs | 100 | | maxServerGroups | 10 | | maxServerGroupMembers | 10 | | maxTotalFloatingIps | -1 | | maxSecurityGroups | -1 | | maxSecurityGroupRules | -1 | | totalRAMUsed | 606208 | | totalCoresUsed | 148 | | totalInstancesUsed | 49 | | totalFloatingIpsUsed | 0 | | totalSecurityGroupsUsed | 0 | | totalServerGroupsUsed | 0 | | maxTotalVolumes | -1 | | maxTotalSnapshots | 10 | | maxTotalVolumeGigabytes | 4096 | | maxTotalBackups | 10 | | maxTotalBackupGigabytes | 1000 | | totalVolumesUsed | 0 | | totalGigabytesUsed | 0 | | totalSnapshotsUsed | 0 | | totalBackupsUsed | 0 | | totalBackupGigabytesUsed | 0 | +--------------------------+---------+ + pushd /opt/ciman/openstack-hot /opt/ciman/openstack-hot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium + lftools openstack --os-cloud vex stack create releng-openflowplugin-csit-3node-clustering-only-calcium-227 csit-2-instance-type.yaml /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/stack-parameters.yaml Creating stack releng-openflowplugin-csit-3node-clustering-only-calcium-227 Waiting to initialize infrastructure... Stack initialization successful. ------------------------------------ Stack Details ------------------------------------ {'added': None, 'capabilities': [], 'created_at': '2024-04-22T04:12:23Z', 'deleted': None, 'deleted_at': None, 'description': 'No description', 'environment': None, 'environment_files': None, 'files': None, 'files_container': None, 'id': 'bfb7be0a-783a-4db0-9002-72aebdface40', 'is_rollback_disabled': True, 'links': [{'href': 'https://orchestration.public.mtl1.vexxhost.net/v1/12c36e260d8e4bb2913965203b1b491f/stacks/releng-openflowplugin-csit-3node-clustering-only-calcium-227/bfb7be0a-783a-4db0-9002-72aebdface40', 'rel': 'self'}], 'location': Munch({'cloud': 'vex', 'region_name': 'ca-ymq-1', 'zone': None, 'project': Munch({'id': '12c36e260d8e4bb2913965203b1b491f', 'name': '61975f2c-7c17-4d69-82fa-c3ae420ad6fd', 'domain_id': None, 'domain_name': 'Default'})}), 'name': 'releng-openflowplugin-csit-3node-clustering-only-calcium-227', 'notification_topics': [], 'outputs': [{'description': 'IP addresses of the 2nd vm types', 'output_key': 'vm_1_ips', 'output_value': ['10.30.170.53']}, {'description': 'IP addresses of the 1st vm types', 'output_key': 'vm_0_ips', 'output_value': ['10.30.170.145', '10.30.170.65', '10.30.170.189']}], 'owner_id': ****, 'parameters': {'OS::project_id': '12c36e260d8e4bb2913965203b1b491f', 'OS::stack_id': 'bfb7be0a-783a-4db0-9002-72aebdface40', 'OS::stack_name': 'releng-openflowplugin-csit-3node-clustering-only-calcium-227', 'job_name': '57946-227', 'silo': 'releng', 'vm_0_count': '3', 'vm_0_flavor': 'v3-standard-4', 'vm_0_image': 'ZZCI - CentOS Stream 8 - builder - x86_64 - ' '20240401-160147.446', 'vm_1_count': '1', 'vm_1_flavor': 'v3-standard-2', 'vm_1_image': 'ZZCI - Ubuntu 18.04 - mininet-ovs-28 - x86_64 - ' '20230601-180106.003'}, 'parent_id': None, 'replaced': None, 'status': 'CREATE_COMPLETE', 'status_reason': 'Stack CREATE completed successfully', 'tags': [], 'template': None, 'template_description': 'No description', 'template_url': None, 'timeout_mins': 15, 'unchanged': None, 'updated': None, 'updated_at': None, 'user_project_id': 'c1828abb386e42c3b1f02a22e2957504'} ------------------------------------ + popd /w/workspace/openflowplugin-csit-3node-clustering-only-calcium [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash -l /tmp/jenkins17462288333908076572.sh ---> Copy SSH public keys to CSIT lab Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools[openstack] kubernetes python-heatclient python-openstackclient lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH Warning: Permanently added '10.30.170.53' (ECDSA) to the list of known hosts. SSH not responding on 10.30.170.65. Retrying in 10 seconds... SSH not responding on 10.30.170.145. Retrying in 10 seconds... releng-57946-227-1-mininet-ovs-28-0 Successfully copied public keys to slave 10.30.170.53 Process 6699 ready. Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. 10.30.170.189 releng-57946-227-0-builder-2.novalocal Successfully copied public keys to slave 10.30.170.189 Ping to 10.30.170.65 successful. Ping to 10.30.170.145 successful. Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. 10.30.170.145 releng-57946-227-0-builder-0.novalocal Successfully copied public keys to slave 10.30.170.145 Process 6700 ready. 10.30.170.65 releng-57946-227-0-builder-1.novalocal Successfully copied public keys to slave 10.30.170.65 Process 6701 ready. Process 6702 ready. SSH ready on all stack servers. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash -l /tmp/jenkins15199214220787095348.sh Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Creating python3 venv at /tmp/venv-WFZV lf-activate-venv(): INFO: Save venv in file: /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.robot_venv lf-activate-venv(): INFO: Installing: setuptools wheel lf-activate-venv(): INFO: Adding /tmp/venv-WFZV/bin to PATH + echo 'Installing Python Requirements' Installing Python Requirements + cat + python -m pip install -r requirements.txt Looking in indexes: https://nexus3.opendaylight.org/repository/PyPi/simple Collecting docker-py (from -r requirements.txt (line 1)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/docker-py/1.10.6/docker_py-1.10.6-py2.py3-none-any.whl (50 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 50.0/50.0 kB 8.5 MB/s eta 0:00:00 Collecting ipaddr (from -r requirements.txt (line 2)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/ipaddr/2.2.0/ipaddr-2.2.0.tar.gz (26 kB) Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Collecting netaddr (from -r requirements.txt (line 3)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/netaddr/1.2.1/netaddr-1.2.1-py3-none-any.whl (2.3 MB) Collecting netifaces (from -r requirements.txt (line 4)) Using cached netifaces-0.11.0-cp311-cp311-linux_x86_64.whl Collecting pyhocon (from -r requirements.txt (line 5)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/pyhocon/0.3.60/pyhocon-0.3.60.tar.gz (158 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 158.8/158.8 kB 24.9 MB/s eta 0:00:00 Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Collecting requests (from -r requirements.txt (line 6)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/requests/2.31.0/requests-2.31.0-py3-none-any.whl (62 kB) Collecting robotframework (from -r requirements.txt (line 7)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework/7.0/robotframework-7.0-py3-none-any.whl (726 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 726.3/726.3 kB 17.1 MB/s eta 0:00:00 Collecting robotframework-httplibrary (from -r requirements.txt (line 8)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-httplibrary/0.4.2/robotframework-httplibrary-0.4.2.tar.gz (9.1 kB) Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Collecting robotframework-requests==0.9.3 (from -r requirements.txt (line 9)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-requests/0.9.3/robotframework_requests-0.9.3-py3-none-any.whl (21 kB) Collecting robotframework-selenium2library (from -r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-selenium2library/3.0.0/robotframework_selenium2library-3.0.0-py2.py3-none-any.whl (6.2 kB) Collecting robotframework-sshlibrary==3.8.0 (from -r requirements.txt (line 11)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-sshlibrary/3.8.0/robotframework-sshlibrary-3.8.0.tar.gz (51 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 51.4/51.4 kB 10.6 MB/s eta 0:00:00 Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Collecting scapy (from -r requirements.txt (line 12)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/scapy/2.5.0/scapy-2.5.0.tar.gz (1.3 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.3/1.3 MB 18.3 MB/s eta 0:00:00 Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Collecting jsonpath-rw (from -r requirements.txt (line 15)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/jsonpath-rw/1.4.0/jsonpath-rw-1.4.0.tar.gz (13 kB) Preparing metadata (setup.py): started Preparing metadata (setup.py): finished with status 'done' Collecting elasticsearch (from -r requirements.txt (line 18)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch/8.13.0/elasticsearch-8.13.0-py3-none-any.whl (451 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 451.0/451.0 kB 21.0 MB/s eta 0:00:00 Collecting elasticsearch-dsl (from -r requirements.txt (line 19)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch-dsl/8.13.0/elasticsearch_dsl-8.13.0-py3-none-any.whl (88 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 88.6/88.6 kB 11.6 MB/s eta 0:00:00 Collecting pyangbind (from -r requirements.txt (line 22)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/pyangbind/0.8.5/pyangbind-0.8.5-py3-none-any.whl (52 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 52.9/52.9 kB 9.9 MB/s eta 0:00:00 Collecting isodate (from -r requirements.txt (line 25)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/isodate/0.6.1/isodate-0.6.1-py2.py3-none-any.whl (41 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 41.7/41.7 kB 8.1 MB/s eta 0:00:00 Collecting jmespath (from -r requirements.txt (line 28)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/jmespath/1.0.1/jmespath-1.0.1-py3-none-any.whl (20 kB) Collecting jsonpatch (from -r requirements.txt (line 31)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/jsonpatch/1.33/jsonpatch-1.33-py2.py3-none-any.whl (12 kB) Collecting paramiko>=1.15.3 (from robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/paramiko/3.4.0/paramiko-3.4.0-py3-none-any.whl (225 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 225.9/225.9 kB 11.7 MB/s eta 0:00:00 Collecting scp>=0.13.0 (from robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/scp/0.14.5/scp-0.14.5-py2.py3-none-any.whl (8.7 kB) Collecting docker-pycreds>=0.2.1 (from docker-py->-r requirements.txt (line 1)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/docker-pycreds/0.4.0/docker_pycreds-0.4.0-py2.py3-none-any.whl (9.0 kB) Collecting six>=1.4.0 (from docker-py->-r requirements.txt (line 1)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/six/1.16.0/six-1.16.0-py2.py3-none-any.whl (11 kB) Collecting websocket-client>=0.32.0 (from docker-py->-r requirements.txt (line 1)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/websocket-client/1.7.0/websocket_client-1.7.0-py3-none-any.whl (58 kB) Collecting pyparsing<4,>=2 (from pyhocon->-r requirements.txt (line 5)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/pyparsing/3.1.2/pyparsing-3.1.2-py3-none-any.whl (103 kB) Collecting charset-normalizer<4,>=2 (from requests->-r requirements.txt (line 6)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/charset-normalizer/3.3.2/charset_normalizer-3.3.2-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (140 kB) Collecting idna<4,>=2.5 (from requests->-r requirements.txt (line 6)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/idna/3.7/idna-3.7-py3-none-any.whl (66 kB) Collecting urllib3<3,>=1.21.1 (from requests->-r requirements.txt (line 6)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/urllib3/2.2.1/urllib3-2.2.1-py3-none-any.whl (121 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 121.1/121.1 kB 18.6 MB/s eta 0:00:00 Collecting certifi>=2017.4.17 (from requests->-r requirements.txt (line 6)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/certifi/2024.2.2/certifi-2024.2.2-py3-none-any.whl (163 kB) Collecting webtest>=2.0 (from robotframework-httplibrary->-r requirements.txt (line 8)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/webtest/3.0.0/WebTest-3.0.0-py3-none-any.whl (31 kB) Collecting jsonpointer (from robotframework-httplibrary->-r requirements.txt (line 8)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/jsonpointer/2.4/jsonpointer-2.4-py2.py3-none-any.whl (7.8 kB) Collecting robotframework-seleniumlibrary>=3.0.0 (from robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-seleniumlibrary/6.3.0/robotframework_seleniumlibrary-6.3.0-py2.py3-none-any.whl (97 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 97.6/97.6 kB 2.3 MB/s eta 0:00:00 Collecting ply (from jsonpath-rw->-r requirements.txt (line 15)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/ply/3.11/ply-3.11-py2.py3-none-any.whl (49 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 49.6/49.6 kB 10.0 MB/s eta 0:00:00 Collecting decorator (from jsonpath-rw->-r requirements.txt (line 15)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/decorator/5.1.1/decorator-5.1.1-py3-none-any.whl (9.1 kB) Collecting elastic-transport<9,>=8.13 (from elasticsearch->-r requirements.txt (line 18)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elastic-transport/8.13.0/elastic_transport-8.13.0-py3-none-any.whl (64 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 64.3/64.3 kB 8.2 MB/s eta 0:00:00 Collecting python-dateutil (from elasticsearch-dsl->-r requirements.txt (line 19)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/python-dateutil/2.9.0.post0/python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB) Collecting pyang (from pyangbind->-r requirements.txt (line 22)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/pyang/2.6.0/pyang-2.6.0-py2.py3-none-any.whl (594 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 594.1/594.1 kB 30.5 MB/s eta 0:00:00 Collecting lxml (from pyangbind->-r requirements.txt (line 22)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/lxml/5.2.1/lxml-5.2.1-cp311-cp311-manylinux_2_28_x86_64.whl (5.0 MB) Collecting regex (from pyangbind->-r requirements.txt (line 22)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/regex/2024.4.16/regex-2024.4.16-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (785 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 785.1/785.1 kB 10.2 MB/s eta 0:00:00 Collecting enum34 (from pyangbind->-r requirements.txt (line 22)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/enum34/1.1.10/enum34-1.1.10-py3-none-any.whl (11 kB) Collecting bcrypt>=3.2 (from paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/bcrypt/4.1.2/bcrypt-4.1.2-cp39-abi3-manylinux_2_28_x86_64.whl (698 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 698.9/698.9 kB 13.5 MB/s eta 0:00:00 Collecting cryptography>=3.3 (from paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/cryptography/42.0.5/cryptography-42.0.5-cp39-abi3-manylinux_2_28_x86_64.whl (4.6 MB) Collecting pynacl>=1.5 (from paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/pynacl/1.5.0/PyNaCl-1.5.0-cp36-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.manylinux_2_24_x86_64.whl (856 kB) Collecting selenium>=4.3.0 (from robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/selenium/4.19.0/selenium-4.19.0-py3-none-any.whl (10.5 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 10.5/10.5 MB 46.0 MB/s eta 0:00:00 Collecting robotframework-pythonlibcore>=3.0.0 (from robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-pythonlibcore/4.4.1/robotframework_pythonlibcore-4.4.1-py2.py3-none-any.whl (12 kB) Collecting WebOb>=1.2 (from webtest>=2.0->robotframework-httplibrary->-r requirements.txt (line 8)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/webob/1.8.7/WebOb-1.8.7-py2.py3-none-any.whl (114 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 115.0/115.0 kB 20.8 MB/s eta 0:00:00 Collecting waitress>=0.8.5 (from webtest>=2.0->robotframework-httplibrary->-r requirements.txt (line 8)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/waitress/3.0.0/waitress-3.0.0-py3-none-any.whl (56 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 56.7/56.7 kB 6.7 MB/s eta 0:00:00 Collecting beautifulsoup4 (from webtest>=2.0->robotframework-httplibrary->-r requirements.txt (line 8)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/beautifulsoup4/4.12.3/beautifulsoup4-4.12.3-py3-none-any.whl (147 kB) Collecting cffi>=1.12 (from cryptography>=3.3->paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/cffi/1.16.0/cffi-1.16.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (464 kB) Collecting trio~=0.17 (from selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/trio/0.25.0/trio-0.25.0-py3-none-any.whl (467 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 467.2/467.2 kB 9.6 MB/s eta 0:00:00 Collecting trio-websocket~=0.9 (from selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/trio-websocket/0.11.1/trio_websocket-0.11.1-py3-none-any.whl (17 kB) Collecting typing_extensions>=4.9.0 (from selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/typing-extensions/4.11.0/typing_extensions-4.11.0-py3-none-any.whl (34 kB) Collecting soupsieve>1.2 (from beautifulsoup4->webtest>=2.0->robotframework-httplibrary->-r requirements.txt (line 8)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/soupsieve/2.5/soupsieve-2.5-py3-none-any.whl (36 kB) Collecting pycparser (from cffi>=1.12->cryptography>=3.3->paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/pycparser/2.22/pycparser-2.22-py3-none-any.whl (117 kB) Collecting attrs>=23.2.0 (from trio~=0.17->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/attrs/23.2.0/attrs-23.2.0-py3-none-any.whl (60 kB) Collecting sortedcontainers (from trio~=0.17->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/sortedcontainers/2.4.0/sortedcontainers-2.4.0-py2.py3-none-any.whl (29 kB) Collecting outcome (from trio~=0.17->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/outcome/1.3.0.post0/outcome-1.3.0.post0-py2.py3-none-any.whl (10 kB) Collecting sniffio>=1.3.0 (from trio~=0.17->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/sniffio/1.3.1/sniffio-1.3.1-py3-none-any.whl (10 kB) Collecting wsproto>=0.14 (from trio-websocket~=0.9->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/wsproto/1.2.0/wsproto-1.2.0-py3-none-any.whl (24 kB) Collecting pysocks!=1.5.7,<2.0,>=1.5.6 (from urllib3[socks]<3,>=1.26->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/pysocks/1.7.1/PySocks-1.7.1-py3-none-any.whl (16 kB) Collecting h11<1,>=0.9.0 (from wsproto>=0.14->trio-websocket~=0.9->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10)) Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/h11/0.14.0/h11-0.14.0-py3-none-any.whl (58 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 58.3/58.3 kB 8.6 MB/s eta 0:00:00 Building wheels for collected packages: robotframework-sshlibrary, ipaddr, pyhocon, robotframework-httplibrary, scapy, jsonpath-rw Building wheel for robotframework-sshlibrary (setup.py): started Building wheel for robotframework-sshlibrary (setup.py): finished with status 'done' Created wheel for robotframework-sshlibrary: filename=robotframework_sshlibrary-3.8.0-py3-none-any.whl size=55119 sha256=6d40f7d5392dca8d57cd87a92b835c2556e48057cb0818ba086003e5db97ca15 Stored in directory: /home/jenkins/.cache/pip/wheels/f7/c9/b3/a977b7bcc410d45ae27d240df3d00a12585509180e373ecccc Building wheel for ipaddr (setup.py): started Building wheel for ipaddr (setup.py): finished with status 'done' Created wheel for ipaddr: filename=ipaddr-2.2.0-py3-none-any.whl size=18280 sha256=6211e273a56739e685d251662d60b5ecbd63e7d08e2b1abafc6e9bd19b14d2a6 Stored in directory: /home/jenkins/.cache/pip/wheels/dc/6c/04/da2d847fa8d45c59af3e1d83e2acc29cb8adcbaf04c0898dbf Building wheel for pyhocon (setup.py): started Building wheel for pyhocon (setup.py): finished with status 'done' Created wheel for pyhocon: filename=pyhocon-0.3.60-py3-none-any.whl size=20865 sha256=023fcd30aa310f81e607ae268c377f0053283e7bd41eca6dc08c105f103aaceb Stored in directory: /home/jenkins/.cache/pip/wheels/02/09/4e/04a983205b51f96912ae6fc13f56f3050a586c98200fe36725 Building wheel for robotframework-httplibrary (setup.py): started Building wheel for robotframework-httplibrary (setup.py): finished with status 'done' Created wheel for robotframework-httplibrary: filename=robotframework_httplibrary-0.4.2-py3-none-any.whl size=9953 sha256=b5129e36847e95ec7962dc379df41e16529070d7c72f94a501161b7f22512192 Stored in directory: /home/jenkins/.cache/pip/wheels/aa/bc/0d/9a20dd51effef392aae2733cb4c7b66c6fa29fca33d88b57ed Building wheel for scapy (setup.py): started Building wheel for scapy (setup.py): finished with status 'done' Created wheel for scapy: filename=scapy-2.5.0-py2.py3-none-any.whl size=1444327 sha256=0ff3a9838d4b23ff60514411180355023bf7e41e9eab0d4213f7644b9bf8db08 Stored in directory: /home/jenkins/.cache/pip/wheels/3b/a7/d6/87eeec31fbfcb2bd8ace7dad8172b6d6ebbe1ec6827dbb389a Building wheel for jsonpath-rw (setup.py): started Building wheel for jsonpath-rw (setup.py): finished with status 'done' Created wheel for jsonpath-rw: filename=jsonpath_rw-1.4.0-py3-none-any.whl size=15130 sha256=aa124986ee36f277015153f6174e87f4cbf7253229edd52e713e6ff950b7d054 Stored in directory: /home/jenkins/.cache/pip/wheels/f1/54/63/9a8da38cefae13755097b36cc852decc25d8ef69c37d58d4eb Successfully built robotframework-sshlibrary ipaddr pyhocon robotframework-httplibrary scapy jsonpath-rw Installing collected packages: sortedcontainers, ply, netifaces, ipaddr, enum34, websocket-client, WebOb, waitress, urllib3, typing_extensions, soupsieve, sniffio, six, scapy, robotframework-pythonlibcore, robotframework, regex, pysocks, pyparsing, pycparser, netaddr, lxml, jsonpointer, jmespath, idna, h11, decorator, charset-normalizer, certifi, bcrypt, attrs, wsproto, requests, python-dateutil, pyhocon, pyang, outcome, jsonpath-rw, jsonpatch, isodate, elastic-transport, docker-pycreds, cffi, beautifulsoup4, webtest, trio, robotframework-requests, pynacl, pyangbind, elasticsearch, docker-py, cryptography, trio-websocket, robotframework-httplibrary, paramiko, elasticsearch-dsl, selenium, scp, robotframework-sshlibrary, robotframework-seleniumlibrary, robotframework-selenium2library Successfully installed WebOb-1.8.7 attrs-23.2.0 bcrypt-4.1.2 beautifulsoup4-4.12.3 certifi-2024.2.2 cffi-1.16.0 charset-normalizer-3.3.2 cryptography-42.0.5 decorator-5.1.1 docker-py-1.10.6 docker-pycreds-0.4.0 elastic-transport-8.13.0 elasticsearch-8.13.0 elasticsearch-dsl-8.13.0 enum34-1.1.10 h11-0.14.0 idna-3.7 ipaddr-2.2.0 isodate-0.6.1 jmespath-1.0.1 jsonpatch-1.33 jsonpath-rw-1.4.0 jsonpointer-2.4 lxml-5.2.1 netaddr-1.2.1 netifaces-0.11.0 outcome-1.3.0.post0 paramiko-3.4.0 ply-3.11 pyang-2.6.0 pyangbind-0.8.5 pycparser-2.22 pyhocon-0.3.60 pynacl-1.5.0 pyparsing-3.1.2 pysocks-1.7.1 python-dateutil-2.9.0.post0 regex-2024.4.16 requests-2.31.0 robotframework-7.0 robotframework-httplibrary-0.4.2 robotframework-pythonlibcore-4.4.1 robotframework-requests-0.9.3 robotframework-selenium2library-3.0.0 robotframework-seleniumlibrary-6.3.0 robotframework-sshlibrary-3.8.0 scapy-2.5.0 scp-0.14.5 selenium-4.19.0 six-1.16.0 sniffio-1.3.1 sortedcontainers-2.4.0 soupsieve-2.5 trio-0.25.0 trio-websocket-0.11.1 typing_extensions-4.11.0 urllib3-2.2.1 waitress-3.0.0 websocket-client-1.7.0 webtest-3.0.0 wsproto-1.2.0 + pip freeze attrs==23.2.0 bcrypt==4.1.2 beautifulsoup4==4.12.3 certifi==2024.2.2 cffi==1.16.0 charset-normalizer==3.3.2 cryptography==42.0.5 decorator==5.1.1 distlib==0.3.8 docker-py==1.10.6 docker-pycreds==0.4.0 elastic-transport==8.13.0 elasticsearch==8.13.0 elasticsearch-dsl==8.13.0 enum34==1.1.10 filelock==3.13.4 h11==0.14.0 idna==3.7 ipaddr==2.2.0 isodate==0.6.1 jmespath==1.0.1 jsonpatch==1.33 jsonpath-rw==1.4.0 jsonpointer==2.4 lxml==5.2.1 netaddr==1.2.1 netifaces==0.11.0 outcome==1.3.0.post0 paramiko==3.4.0 platformdirs==4.2.0 ply==3.11 pyang==2.6.0 pyangbind==0.8.5 pycparser==2.22 pyhocon==0.3.60 PyNaCl==1.5.0 pyparsing==3.1.2 PySocks==1.7.1 python-dateutil==2.9.0.post0 regex==2024.4.16 requests==2.31.0 robotframework==7.0 robotframework-httplibrary==0.4.2 robotframework-pythonlibcore==4.4.1 robotframework-requests==0.9.3 robotframework-selenium2library==3.0.0 robotframework-seleniumlibrary==6.3.0 robotframework-sshlibrary==3.8.0 scapy==2.5.0 scp==0.14.5 selenium==4.19.0 six==1.16.0 sniffio==1.3.1 sortedcontainers==2.4.0 soupsieve==2.5 trio==0.25.0 trio-websocket==0.11.1 typing_extensions==4.11.0 urllib3==2.2.1 virtualenv==20.25.3 waitress==3.0.0 WebOb==1.8.7 websocket-client==1.7.0 WebTest==3.0.0 wsproto==1.2.0 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path 'env.properties' [EnvInject] - Variables injected successfully. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash -l /tmp/jenkins11069051882937550894.sh Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: python-heatclient python-openstackclient yq ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts. lftools 0.37.10 requires urllib3<2.1.0, but you have urllib3 2.2.1 which is incompatible. lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH + ODL_SYSTEM=() + TOOLS_SYSTEM=() + OPENSTACK_SYSTEM=() + OPENSTACK_CONTROLLERS=() + mapfile -t ADDR ++ jq -r '.outputs[] | select(.output_key | match("^vm_[0-9]+_ips$")) | .output_value | .[]' ++ openstack stack show -f json -c outputs releng-openflowplugin-csit-3node-clustering-only-calcium-227 + for i in "${ADDR[@]}" ++ ssh 10.30.170.53 hostname -s Warning: Permanently added '10.30.170.53' (ECDSA) to the list of known hosts. + REMHOST=releng-57946-227-1-mininet-ovs-28-0 + case ${REMHOST} in + TOOLS_SYSTEM=("${TOOLS_SYSTEM[@]}" "${i}") + for i in "${ADDR[@]}" ++ ssh 10.30.170.145 hostname -s Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + REMHOST=releng-57946-227-0-builder-0 + case ${REMHOST} in + ODL_SYSTEM=("${ODL_SYSTEM[@]}" "${i}") + for i in "${ADDR[@]}" ++ ssh 10.30.170.65 hostname -s Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + REMHOST=releng-57946-227-0-builder-1 + case ${REMHOST} in + ODL_SYSTEM=("${ODL_SYSTEM[@]}" "${i}") + for i in "${ADDR[@]}" ++ ssh 10.30.170.189 hostname -s Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + REMHOST=releng-57946-227-0-builder-2 + case ${REMHOST} in + ODL_SYSTEM=("${ODL_SYSTEM[@]}" "${i}") + echo NUM_ODL_SYSTEM=3 + echo NUM_TOOLS_SYSTEM=1 + '[' '' == yes ']' + NUM_OPENSTACK_SYSTEM=0 + echo NUM_OPENSTACK_SYSTEM=0 + '[' 0 -eq 2 ']' + echo ODL_SYSTEM_IP=10.30.170.145 ++ seq 0 2 + for i in $(seq 0 $(( ${#ODL_SYSTEM[@]} - 1 ))) + echo ODL_SYSTEM_1_IP=10.30.170.145 + for i in $(seq 0 $(( ${#ODL_SYSTEM[@]} - 1 ))) + echo ODL_SYSTEM_2_IP=10.30.170.65 + for i in $(seq 0 $(( ${#ODL_SYSTEM[@]} - 1 ))) + echo ODL_SYSTEM_3_IP=10.30.170.189 + echo TOOLS_SYSTEM_IP=10.30.170.53 ++ seq 0 0 + for i in $(seq 0 $(( ${#TOOLS_SYSTEM[@]} - 1 ))) + echo TOOLS_SYSTEM_1_IP=10.30.170.53 + openstack_index=0 + NUM_OPENSTACK_CONTROL_NODES=1 + echo NUM_OPENSTACK_CONTROL_NODES=1 ++ seq 0 0 + for i in $(seq 0 $((NUM_OPENSTACK_CONTROL_NODES - 1))) + echo OPENSTACK_CONTROL_NODE_1_IP= + NUM_OPENSTACK_COMPUTE_NODES=-1 + echo NUM_OPENSTACK_COMPUTE_NODES=-1 + '[' -1 -ge 2 ']' ++ seq 0 -2 + NUM_OPENSTACK_HAPROXY_NODES=0 + echo NUM_OPENSTACK_HAPROXY_NODES=0 ++ seq 0 -1 + echo 'Contents of slave_addresses.txt:' Contents of slave_addresses.txt: + cat slave_addresses.txt NUM_ODL_SYSTEM=3 NUM_TOOLS_SYSTEM=1 NUM_OPENSTACK_SYSTEM=0 ODL_SYSTEM_IP=10.30.170.145 ODL_SYSTEM_1_IP=10.30.170.145 ODL_SYSTEM_2_IP=10.30.170.65 ODL_SYSTEM_3_IP=10.30.170.189 TOOLS_SYSTEM_IP=10.30.170.53 TOOLS_SYSTEM_1_IP=10.30.170.53 NUM_OPENSTACK_CONTROL_NODES=1 OPENSTACK_CONTROL_NODE_1_IP= NUM_OPENSTACK_COMPUTE_NODES=-1 NUM_OPENSTACK_HAPROXY_NODES=0 [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path 'slave_addresses.txt' [EnvInject] - Variables injected successfully. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/sh /tmp/jenkins7617288721394659778.sh Preparing for JRE Version 17 Karaf artifact is karaf Karaf project is integration Java home is /usr/lib/jvm/java-17-openjdk [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path 'set_variables.env' [EnvInject] - Variables injected successfully. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins14007125196470661373.sh Distribution bundle URL is https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip Distribution bundle is karaf-0.21.0.zip Distribution bundle version is 0.21.0 Distribution folder is karaf-0.21.0 Nexus prefix is https://nexus.opendaylight.org [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties file path 'detect_variables.env' [EnvInject] - Variables injected successfully. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash -l /tmp/jenkins14773039737323630766.sh Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: python-heatclient python-openstackclient lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH Copying common-functions.sh to /tmp Copying common-functions.sh to 10.30.170.53:/tmp Warning: Permanently added '10.30.170.53' (ECDSA) to the list of known hosts. Copying common-functions.sh to 10.30.170.145:/tmp Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. Copying common-functions.sh to 10.30.170.65:/tmp Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. Copying common-functions.sh to 10.30.170.189:/tmp Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins9289616419466672157.sh common-functions.sh is being sourced common-functions environment: MAVENCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ACTUALFEATURES: FEATURESCONF: /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg CUSTOMPROP: /tmp/karaf-0.21.0/etc/custom.properties LOGCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg MEMCONF: /tmp/karaf-0.21.0/bin/setenv CONTROLLERMEM: 2048m AKKACONF: /tmp/karaf-0.21.0/configuration/initial/akka.conf MODULESCONF: /tmp/karaf-0.21.0/configuration/initial/modules.conf MODULESHARDSCONF: /tmp/karaf-0.21.0/configuration/initial/module-shards.conf SUITES: ################################################# ## Configure Cluster and Start ## ################################################# ACTUALFEATURES: odl-infrautils-ready,odl-jolokia,odl-openflowplugin-flow-services-rest,odl-openflowplugin-app-table-miss-enforcer SPACE_SEPARATED_FEATURES: odl-infrautils-ready odl-jolokia odl-openflowplugin-flow-services-rest odl-openflowplugin-app-table-miss-enforcer Locating script plan to use... Finished running script plans Configuring member-1 with IP address 10.30.170.145 Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + source /tmp/common-functions.sh karaf-0.21.0 ++ [[ /tmp/common-functions.sh == \/\t\m\p\/\c\o\n\f\i\g\u\r\a\t\i\o\n\-\s\c\r\i\p\t\.\s\h ]] common-functions.sh is being sourced ++ echo 'common-functions.sh is being sourced' ++ BUNDLEFOLDER=karaf-0.21.0 ++ export MAVENCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ++ MAVENCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ++ export FEATURESCONF=/tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ++ FEATURESCONF=/tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ++ export CUSTOMPROP=/tmp/karaf-0.21.0/etc/custom.properties ++ CUSTOMPROP=/tmp/karaf-0.21.0/etc/custom.properties ++ export LOGCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ++ LOGCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ++ export MEMCONF=/tmp/karaf-0.21.0/bin/setenv ++ MEMCONF=/tmp/karaf-0.21.0/bin/setenv ++ export CONTROLLERMEM= ++ CONTROLLERMEM= ++ export AKKACONF=/tmp/karaf-0.21.0/configuration/initial/akka.conf ++ AKKACONF=/tmp/karaf-0.21.0/configuration/initial/akka.conf ++ export MODULESCONF=/tmp/karaf-0.21.0/configuration/initial/modules.conf ++ MODULESCONF=/tmp/karaf-0.21.0/configuration/initial/modules.conf ++ export MODULESHARDSCONF=/tmp/karaf-0.21.0/configuration/initial/module-shards.conf ++ MODULESHARDSCONF=/tmp/karaf-0.21.0/configuration/initial/module-shards.conf ++ print_common_env ++ cat common-functions environment: MAVENCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ACTUALFEATURES: FEATURESCONF: /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg CUSTOMPROP: /tmp/karaf-0.21.0/etc/custom.properties LOGCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg MEMCONF: /tmp/karaf-0.21.0/bin/setenv CONTROLLERMEM: AKKACONF: /tmp/karaf-0.21.0/configuration/initial/akka.conf MODULESCONF: /tmp/karaf-0.21.0/configuration/initial/modules.conf MODULESHARDSCONF: /tmp/karaf-0.21.0/configuration/initial/module-shards.conf SUITES: ++ SSH='ssh -t -t' ++ extra_services_cntl=' dnsmasq.service httpd.service libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service rabbitmq-server.service ' ++ extra_services_cmp=' libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service ' Changing to /tmp Downloading the distribution from https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip + echo 'Changing to /tmp' + cd /tmp + echo 'Downloading the distribution from https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip' + wget --progress=dot:mega https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip --2024-04-22 04:14:22-- https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip Resolving nexus.opendaylight.org (nexus.opendaylight.org)... 199.204.45.87, 2604:e100:1:0:f816:3eff:fe45:48d6 Connecting to nexus.opendaylight.org (nexus.opendaylight.org)|199.204.45.87|:443... connected. HTTP request sent, awaiting response... 200 OK Length: 290466339 (277M) [application/zip] Saving to: ‘karaf-0.21.0.zip’ 0K ........ ........ ........ ........ ........ ........ 1% 71.6M 4s 3072K ........ ........ ........ ........ ........ ........ 2% 126M 3s 6144K ........ ........ ........ ........ ........ ........ 3% 120M 3s 9216K ........ ........ ........ ........ ........ ........ 4% 27.5M 4s 12288K ........ ........ ........ ........ ........ ........ 5% 166M 4s 15360K ........ ........ ........ ........ ........ ........ 6% 235M 3s 18432K ........ ........ ........ ........ ........ ........ 7% 246M 3s 21504K ........ ........ ........ ........ ........ ........ 8% 275M 3s 24576K ........ ........ ........ ........ ........ ........ 9% 296M 2s 27648K ........ ........ ........ ........ ........ ........ 10% 321M 2s 30720K ........ ........ ........ ........ ........ ........ 11% 100M 2s 33792K ........ ........ ........ ........ ........ ........ 12% 231M 2s 36864K ........ ........ ........ ........ ........ ........ 14% 339M 2s 39936K ........ ........ ........ ........ ........ ........ 15% 305M 2s 43008K ........ ........ ........ ........ ........ ........ 16% 290M 2s 46080K ........ ........ ........ ........ ........ ........ 17% 332M 2s 49152K ........ ........ ........ ........ ........ ........ 18% 293M 2s 52224K ........ ........ ........ ........ ........ ........ 19% 288M 2s 55296K ........ ........ ........ ........ ........ ........ 20% 383M 1s 58368K ........ ........ ........ ........ ........ ........ 21% 342M 1s 61440K ........ ........ ........ ........ ........ ........ 22% 270M 1s 64512K ........ ........ ........ ........ ........ ........ 23% 260M 1s 67584K ........ ........ ........ ........ ........ ........ 24% 242M 1s 70656K ........ ........ ........ ........ ........ ........ 25% 258M 1s 73728K ........ ........ ........ ........ ........ ........ 27% 289M 1s 76800K ........ ........ ........ ........ ........ ........ 28% 337M 1s 79872K ........ ........ ........ ........ ........ ........ 29% 319M 1s 82944K ........ ........ ........ ........ ........ ........ 30% 310M 1s 86016K ........ ........ ........ ........ ........ ........ 31% 276M 1s 89088K ........ ........ ........ ........ ........ ........ 32% 273M 1s 92160K ........ ........ ........ ........ ........ ........ 33% 309M 1s 95232K ........ ........ ........ ........ ........ ........ 34% 310M 1s 98304K ........ ........ ........ ........ ........ ........ 35% 331M 1s 101376K ........ ........ ........ ........ ........ ........ 36% 317M 1s 104448K ........ ........ ........ ........ ........ ........ 37% 261M 1s 107520K ........ ........ ........ ........ ........ ........ 38% 262M 1s 110592K ........ ........ ........ ........ ........ ........ 40% 193M 1s 113664K ........ ........ ........ ........ ........ ........ 41% 350M 1s 116736K ........ ........ ........ ........ ........ ........ 42% 333M 1s 119808K ........ ........ ........ ........ ........ ........ 43% 264M 1s 122880K ........ ........ ........ ........ ........ ........ 44% 246M 1s 125952K ........ ........ ........ ........ ........ ........ 45% 121M 1s 129024K ........ ........ ........ ........ ........ ........ 46% 266M 1s 132096K ........ ........ ........ ........ ........ ........ 47% 310M 1s 135168K ........ ........ ........ ........ ........ ........ 48% 257M 1s 138240K ........ ........ ........ ........ ........ ........ 49% 278M 1s 141312K ........ ........ ........ ........ ........ ........ 50% 331M 1s 144384K ........ ........ ........ ........ ........ ........ 51% 354M 1s 147456K ........ ........ ........ ........ ........ ........ 53% 341M 1s 150528K ........ ........ ........ ........ ........ ........ 54% 341M 1s 153600K ........ ........ ........ ........ ........ ........ 55% 340M 1s 156672K ........ ........ ........ ........ ........ ........ 56% 320M 1s 159744K ........ ........ ........ ........ ........ ........ 57% 211M 1s 162816K ........ ........ ........ ........ ........ ........ 58% 282M 1s 165888K ........ ........ ........ ........ ........ ........ 59% 336M 1s 168960K ........ ........ ........ ........ ........ ........ 60% 334M 1s 172032K ........ ........ ........ ........ ........ ........ 61% 317M 0s 175104K ........ ........ ........ ........ ........ ........ 62% 342M 0s 178176K ........ ........ ........ ........ ........ ........ 63% 326M 0s 181248K ........ ........ ........ ........ ........ ........ 64% 328M 0s 184320K ........ ........ ........ ........ ........ ........ 66% 333M 0s 187392K ........ ........ ........ ........ ........ ........ 67% 340M 0s 190464K ........ ........ ........ ........ ........ ........ 68% 330M 0s 193536K ........ ........ ........ ........ ........ ........ 69% 336M 0s 196608K ........ ........ ........ ........ ........ ........ 70% 277M 0s 199680K ........ ........ ........ ........ ........ ........ 71% 309M 0s 202752K ........ ........ ........ ........ ........ ........ 72% 308M 0s 205824K ........ ........ ........ ........ ........ ........ 73% 328M 0s 208896K ........ ........ ........ ........ ........ ........ 74% 324M 0s 211968K ........ ........ ........ ........ ........ ........ 75% 310M 0s 215040K ........ ........ ........ ........ ........ ........ 76% 313M 0s 218112K ........ ........ ........ ........ ........ ........ 77% 326M 0s 221184K ........ ........ ........ ........ ........ ........ 79% 322M 0s 224256K ........ ........ ........ ........ ........ ........ 80% 330M 0s 227328K ........ ........ ........ ........ ........ ........ 81% 311M 0s 230400K ........ ........ ........ ........ ........ ........ 82% 308M 0s 233472K ........ ........ ........ ........ ........ ........ 83% 313M 0s 236544K ........ ........ ........ ........ ........ ........ 84% 327M 0s 239616K ........ ........ ........ ........ ........ ........ 85% 326M 0s 242688K ........ ........ ........ ........ ........ ........ 86% 336M 0s 245760K ........ ........ ........ ........ ........ ........ 87% 321M 0s 248832K ........ ........ ........ ........ ........ ........ 88% 325M 0s 251904K ........ ........ ........ ........ ........ ........ 89% 311M 0s 254976K ........ ........ ........ ........ ........ ........ 90% 303M 0s 258048K ........ ........ ........ ........ ........ ........ 92% 309M 0s 261120K ........ ........ ........ ........ ........ ........ 93% 266M 0s 264192K ........ ........ ........ ........ ........ ........ 94% 302M 0s 267264K ........ ........ ........ ........ ........ ........ 95% 387M 0s 270336K ........ ........ ........ ........ ........ ........ 96% 357M 0s 273408K ........ ........ ........ ........ ........ ........ 97% 338M 0s 276480K ........ ........ ........ ........ ........ ........ 98% 336M 0s 279552K ........ ........ ........ ........ ........ ........ 99% 351M 0s 282624K ........ ........ 100% 348M=1.1s 2024-04-22 04:14:23 (248 MB/s) - ‘karaf-0.21.0.zip’ saved [290466339/290466339] Extracting the new controller... + echo 'Extracting the new controller...' + unzip -q karaf-0.21.0.zip Adding external repositories... + echo 'Adding external repositories...' + sed -ie 's%org.ops4j.pax.url.mvn.repositories=%org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases%g' /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg + cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # # If set to true, the following property will not allow any certificate to be used # when accessing Maven repositories through SSL # #org.ops4j.pax.url.mvn.certificateCheck= # # Path to the local Maven settings file. # The repositories defined in this file will be automatically added to the list # of default repositories if the 'org.ops4j.pax.url.mvn.repositories' property # below is not set. # The following locations are checked for the existence of the settings.xml file # * 1. looks for the specified url # * 2. if not found looks for ${user.home}/.m2/settings.xml # * 3. if not found looks for ${maven.home}/conf/settings.xml # * 4. if not found looks for ${M2_HOME}/conf/settings.xml # #org.ops4j.pax.url.mvn.settings= # # Path to the local Maven repository which is used to avoid downloading # artifacts when they already exist locally. # The value of this property will be extracted from the settings.xml file # above, or defaulted to: # System.getProperty( "user.home" ) + "/.m2/repository" # org.ops4j.pax.url.mvn.localRepository=${karaf.home}/${karaf.default.repository} # # Default this to false. It's just weird to use undocumented repos # org.ops4j.pax.url.mvn.useFallbackRepositories=false # # Uncomment if you don't wanna use the proxy settings # from the Maven conf/settings.xml file # # org.ops4j.pax.url.mvn.proxySupport=false # # Comma separated list of repositories scanned when resolving an artifact. # Those repositories will be checked before iterating through the # below list of repositories and even before the local repository # A repository url can be appended with zero or more of the following flags: # @snapshots : the repository contains snaphots # @noreleases : the repository does not contain any released artifacts # # The following property value will add the system folder as a repo. # org.ops4j.pax.url.mvn.defaultRepositories=\ file:${karaf.home}/${karaf.default.repository}@id=system.repository@snapshots,\ file:${karaf.data}/kar@id=kar.repository@multi@snapshots,\ file:${karaf.base}/${karaf.default.repository}@id=child.system.repository@snapshots # Use the default local repo (e.g.~/.m2/repository) as a "remote" repo #org.ops4j.pax.url.mvn.defaultLocalRepoAsRemote=false # # Comma separated list of repositories scanned when resolving an artifact. # The default list includes the following repositories: # http://repo1.maven.org/maven2@id=central # http://repository.springsource.com/maven/bundles/release@id=spring.ebr # http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external # http://zodiac.springsource.com/maven/bundles/release@id=gemini # http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases # https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases # https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases # To add repositories to the default ones, prepend '+' to the list of repositories # to add. # A repository url can be appended with zero or more of the following flags: # @snapshots : the repository contains snapshots # @noreleases : the repository does not contain any released artifacts # @id=repository.id : the id for the repository, just like in the settings.xml this is optional but recommended # org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases ### ^^^ No remote repositories. This is the only ODL change compared to Karaf defaults.Configuring the startup features... + [[ True == \T\r\u\e ]] + echo 'Configuring the startup features...' + sed -ie 's/\(featuresBoot=\|featuresBoot =\)/featuresBoot = odl-infrautils-ready,odl-jolokia,odl-openflowplugin-flow-services-rest,odl-openflowplugin-app-table-miss-enforcer,/g' /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg + FEATURE_TEST_STRING=features-test + FEATURE_TEST_VERSION=0.21.0 + KARAF_VERSION=karaf4 + [[ integration == \i\n\t\e\g\r\a\t\i\o\n ]] + sed -ie 's%\(featuresRepositories=\|featuresRepositories =\)%featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.0/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features,%g' /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg + [[ ! -z '' ]] + cat /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # # Comma separated list of features repositories to register by default # featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.0/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features, file:${karaf.etc}/26fa34cc-b631-49df-84f3-36941131d37f.xml # # Comma separated list of features to install at startup # featuresBoot = odl-infrautils-ready,odl-jolokia,odl-openflowplugin-flow-services-rest,odl-openflowplugin-app-table-miss-enforcer, e5317303-342d-4134-9449-c8c22a0bc5b4 # # Resource repositories (OBR) that the features resolver can use # to resolve requirements/capabilities # # The format of the resourceRepositories is # resourceRepositories=[xml:url|json:url],... # for Instance: # #resourceRepositories=xml:http://host/path/to/index.xml # or #resourceRepositories=json:http://host/path/to/index.json # # # Defines if the boot features are started in asynchronous mode (in a dedicated thread) # featuresBootAsynchronous=false # # Service requirements enforcement # # By default, the feature resolver checks the service requirements/capabilities of # bundles for new features (xml schema >= 1.3.0) in order to automatically installs # the required bundles. # The following flag can have those values: # - disable: service requirements are completely ignored # - default: service requirements are ignored for old features # - enforce: service requirements are always verified # #serviceRequirements=default # # Store cfg file for config element in feature # #configCfgStore=true # # Define if the feature service automatically refresh bundles # autoRefresh=true # # Configuration of features processing mechanism (overrides, blacklisting, modification of features) # XML file defines instructions related to features processing # versions.properties may declare properties to resolve placeholders in XML file # both files are relative to ${karaf.etc} # #featureProcessing=org.apache.karaf.features.xml #featureProcessingVersions=versions.properties + configure_karaf_log karaf4 '' + local -r karaf_version=karaf4 + local -r controllerdebugmap= + local logapi=log4j + grep log4j2 /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n log4j2.rootLogger.level = INFO #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 64MB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE Configuring the karaf log... karaf_version: karaf4, logapi: log4j2 + logapi=log4j2 + echo 'Configuring the karaf log... karaf_version: karaf4, logapi: log4j2' + '[' log4j2 == log4j2 ']' + sed -ie 's/log4j2.appender.rolling.policies.size.size = 64MB/log4j2.appender.rolling.policies.size.size = 1GB/g' /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg + orgmodule=org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver + orgmodule_=org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver + echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN' controllerdebugmap: cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg + echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN' + unset IFS + echo 'controllerdebugmap: ' + '[' -n '' ']' + echo 'cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg' + cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # Common pattern layout for appenders log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n # Root logger log4j2.rootLogger.level = INFO # uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} # Filters for logs marked by org.opendaylight.odlparent.Markers log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL # Loggers configuration # Spifly logger log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN # Security audit logger log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Appenders configuration # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} # Rolling file appender log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i # uncomment to not force a disk flush #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 1GB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 # Audit file appender log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 # OSGi appender log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * # help with identification of maven-related problems with pax-url-aether #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN Configure java home: /usr/lib/jvm/java-17-openjdk max memory: 2048m memconf: /tmp/karaf-0.21.0/bin/setenv + set_java_vars /usr/lib/jvm/java-17-openjdk 2048m /tmp/karaf-0.21.0/bin/setenv + local -r java_home=/usr/lib/jvm/java-17-openjdk + local -r controllermem=2048m + local -r memconf=/tmp/karaf-0.21.0/bin/setenv + echo Configure + echo ' java home: /usr/lib/jvm/java-17-openjdk' + echo ' max memory: 2048m' + echo ' memconf: /tmp/karaf-0.21.0/bin/setenv' + sed -ie 's%^# export JAVA_HOME%export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-17-openjdk}%g' /tmp/karaf-0.21.0/bin/setenv + sed -ie 's/JAVA_MAX_MEM="2048m"/JAVA_MAX_MEM=2048m/g' /tmp/karaf-0.21.0/bin/setenv cat /tmp/karaf-0.21.0/bin/setenv + echo 'cat /tmp/karaf-0.21.0/bin/setenv' + cat /tmp/karaf-0.21.0/bin/setenv #!/bin/sh # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # # handle specific scripts; the SCRIPT_NAME is exactly the name of the Karaf # script: client, instance, shell, start, status, stop, karaf # # if [ "${KARAF_SCRIPT}" == "SCRIPT_NAME" ]; then # Actions go here... # fi # # general settings which should be applied for all scripts go here; please keep # in mind that it is possible that scripts might be executed more than once, e.g. # in example of the start script where the start script is executed first and the # karaf script afterwards. # # # The following section shows the possible configuration options for the default # karaf scripts # export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-17-openjdk} # Location of Java installation # export JAVA_OPTS # Generic JVM options, for instance, where you can pass the memory configuration # export JAVA_NON_DEBUG_OPTS # Additional non-debug JVM options # export EXTRA_JAVA_OPTS # Additional JVM options # export KARAF_HOME # Karaf home folder # export KARAF_DATA # Karaf data folder # export KARAF_BASE # Karaf base folder # export KARAF_ETC # Karaf etc folder # export KARAF_LOG # Karaf log folder # export KARAF_SYSTEM_OPTS # First citizen Karaf options # export KARAF_OPTS # Additional available Karaf options # export KARAF_DEBUG # Enable debug mode # export KARAF_REDIRECT # Enable/set the std/err redirection when using bin/start # export KARAF_NOROOT # Prevent execution as root if set to true Set Java version + echo 'Set Java version' + sudo /usr/sbin/alternatives --install /usr/bin/java java /usr/lib/jvm/java-17-openjdk/bin/java 1 + sudo /usr/sbin/alternatives --set java /usr/lib/jvm/java-17-openjdk/bin/java JDK default version ... + echo 'JDK default version ...' + java -version openjdk version "17.0.6-ea" 2023-01-17 LTS OpenJDK Runtime Environment (Red_Hat-17.0.6.0.9-0.3.ea.el8) (build 17.0.6-ea+9-LTS) OpenJDK 64-Bit Server VM (Red_Hat-17.0.6.0.9-0.3.ea.el8) (build 17.0.6-ea+9-LTS, mixed mode, sharing) Set JAVA_HOME + echo 'Set JAVA_HOME' + export JAVA_HOME=/usr/lib/jvm/java-17-openjdk + JAVA_HOME=/usr/lib/jvm/java-17-openjdk ++ readlink -e /usr/lib/jvm/java-17-openjdk/bin/java Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java Listing all open ports on controller system... + JAVA_RESOLVED=/usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java + echo 'Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java' + echo 'Listing all open ports on controller system...' + netstat -pnatu (Not all processes could be identified, non-owned process info will not be shown, you would have to be root to see it all.) Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN - tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN - tcp 0 0 10.30.170.145:54574 199.204.45.87:443 TIME_WAIT - tcp 0 164 10.30.170.145:22 10.30.171.144:41368 ESTABLISHED - tcp6 0 0 :::111 :::* LISTEN - tcp6 0 0 :::5555 :::* LISTEN - tcp6 0 0 :::22 :::* LISTEN - udp 0 0 10.30.170.145:68 10.30.170.4:67 ESTABLISHED - udp 0 0 0.0.0.0:111 0.0.0.0:* - udp 0 0 127.0.0.1:323 0.0.0.0:* - udp6 0 0 :::111 :::* - udp6 0 0 ::1:323 :::* - Configuring cluster + '[' -f /tmp/custom_shard_config.txt ']' + echo 'Configuring cluster' + /tmp/karaf-0.21.0/bin/configure_cluster.sh 1 10.30.170.145 10.30.170.65 10.30.170.189 ################################################ ## Configure Cluster ## ################################################ NOTE: Cluster configuration files not found. Copying from /tmp/karaf-0.21.0/system/org/opendaylight/controller/sal-clustering-config/9.0.2 Configuring unique name in akka.conf Configuring hostname in akka.conf Configuring data and rpc seed nodes in akka.conf modules = [ { name = "inventory" namespace = "urn:opendaylight:inventory" shard-strategy = "module" }, { name = "topology" namespace = "urn:TBD:params:xml:ns:yang:network-topology" shard-strategy = "module" }, { name = "toaster" namespace = "http://netconfcentral.org/ns/toaster" shard-strategy = "module" } ] Configuring replication type in module-shards.conf ################################################ ## NOTE: Manually restart controller to ## ## apply configuration. ## ################################################ Dump akka.conf + echo 'Dump akka.conf' + cat /tmp/karaf-0.21.0/configuration/initial/akka.conf odl-cluster-data { akka { remote { artery { enabled = on transport = tcp canonical.hostname = "10.30.170.145" canonical.port = 2550 } } cluster { # Using artery. seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.145:2550", "akka://opendaylight-cluster-data@10.30.170.65:2550", "akka://opendaylight-cluster-data@10.30.170.189:2550"] roles = ["member-1"] # when under load we might trip a false positive on the failure detector # failure-detector { # heartbeat-interval = 4 s # acceptable-heartbeat-pause = 16s # } } persistence { # By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by # modifying the following two properties. The directory location specified may be a relative or absolute path. # The relative path is always relative to KARAF_HOME. # snapshot-store.local.dir = "target/snapshots" # Use lz4 compression for LocalSnapshotStore snapshots snapshot-store.local.use-lz4-compression = false # Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB snapshot-store.local.lz4-blocksize = 256KB } disable-default-actor-system-quarantined-event-handling = "false" } } Dump modules.conf + echo 'Dump modules.conf' + cat /tmp/karaf-0.21.0/configuration/initial/modules.conf modules = [ { name = "inventory" namespace = "urn:opendaylight:inventory" shard-strategy = "module" }, { name = "topology" namespace = "urn:TBD:params:xml:ns:yang:network-topology" shard-strategy = "module" }, { name = "toaster" namespace = "http://netconfcentral.org/ns/toaster" shard-strategy = "module" } ] Dump module-shards.conf + echo 'Dump module-shards.conf' + cat /tmp/karaf-0.21.0/configuration/initial/module-shards.conf module-shards = [ { name = "default" shards = [ { name = "default" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "inventory" shards = [ { name="inventory" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "topology" shards = [ { name="topology" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "toaster" shards = [ { name="toaster" replicas = ["member-1", "member-2", "member-3"] } ] } ] Configuring member-2 with IP address 10.30.170.65 Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. common-functions.sh is being sourced + source /tmp/common-functions.sh karaf-0.21.0 ++ [[ /tmp/common-functions.sh == \/\t\m\p\/\c\o\n\f\i\g\u\r\a\t\i\o\n\-\s\c\r\i\p\t\.\s\h ]] ++ echo 'common-functions.sh is being sourced' ++ BUNDLEFOLDER=karaf-0.21.0 ++ export MAVENCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ++ MAVENCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ++ export FEATURESCONF=/tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ++ FEATURESCONF=/tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ++ export CUSTOMPROP=/tmp/karaf-0.21.0/etc/custom.properties ++ CUSTOMPROP=/tmp/karaf-0.21.0/etc/custom.properties ++ export LOGCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ++ LOGCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ++ export MEMCONF=/tmp/karaf-0.21.0/bin/setenv ++ MEMCONF=/tmp/karaf-0.21.0/bin/setenv ++ export CONTROLLERMEM= ++ CONTROLLERMEM= ++ export AKKACONF=/tmp/karaf-0.21.0/configuration/initial/akka.conf ++ AKKACONF=/tmp/karaf-0.21.0/configuration/initial/akka.conf ++ export MODULESCONF=/tmp/karaf-0.21.0/configuration/initial/modules.conf ++ MODULESCONF=/tmp/karaf-0.21.0/configuration/initial/modules.conf ++ export MODULESHARDSCONF=/tmp/karaf-0.21.0/configuration/initial/module-shards.conf ++ MODULESHARDSCONF=/tmp/karaf-0.21.0/configuration/initial/module-shards.conf ++ print_common_env ++ cat common-functions environment: MAVENCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ACTUALFEATURES: FEATURESCONF: /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg CUSTOMPROP: /tmp/karaf-0.21.0/etc/custom.properties LOGCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg MEMCONF: /tmp/karaf-0.21.0/bin/setenv CONTROLLERMEM: AKKACONF: /tmp/karaf-0.21.0/configuration/initial/akka.conf MODULESCONF: /tmp/karaf-0.21.0/configuration/initial/modules.conf MODULESHARDSCONF: /tmp/karaf-0.21.0/configuration/initial/module-shards.conf SUITES: ++ SSH='ssh -t -t' ++ extra_services_cntl=' dnsmasq.service httpd.service libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service rabbitmq-server.service ' ++ extra_services_cmp=' libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service ' Changing to /tmp Downloading the distribution from https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip + echo 'Changing to /tmp' + cd /tmp + echo 'Downloading the distribution from https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip' + wget --progress=dot:mega https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip --2024-04-22 04:14:27-- https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip Resolving nexus.opendaylight.org (nexus.opendaylight.org)... 199.204.45.87, 2604:e100:1:0:f816:3eff:fe45:48d6 Connecting to nexus.opendaylight.org (nexus.opendaylight.org)|199.204.45.87|:443... connected. HTTP request sent, awaiting response... 200 OK Length: 290466339 (277M) [application/zip] Saving to: ‘karaf-0.21.0.zip’ 0K ........ ........ ........ ........ ........ ........ 1% 66.1M 4s 3072K ........ ........ ........ ........ ........ ........ 2% 107M 3s 6144K ........ ........ ........ ........ ........ ........ 3% 138M 3s 9216K ........ ........ ........ ........ ........ ........ 4% 150M 3s 12288K ........ ........ ........ ........ ........ ........ 5% 141M 2s 15360K ........ ........ ........ ........ ........ ........ 6% 181M 2s 18432K ........ ........ ........ ........ ........ ........ 7% 219M 2s 21504K ........ ........ ........ ........ ........ ........ 8% 238M 2s 24576K ........ ........ ........ ........ ........ ........ 9% 245M 2s 27648K ........ ........ ........ ........ ........ ........ 10% 279M 2s 30720K ........ ........ ........ ........ ........ ........ 11% 258M 2s 33792K ........ ........ ........ ........ ........ ........ 12% 326M 1s 36864K ........ ........ ........ ........ ........ ........ 14% 360M 1s 39936K ........ ........ ........ ........ ........ ........ 15% 256M 1s 43008K ........ ........ ........ ........ ........ ........ 16% 265M 1s 46080K ........ ........ ........ ........ ........ ........ 17% 319M 1s 49152K ........ ........ ........ ........ ........ ........ 18% 327M 1s 52224K ........ ........ ........ ........ ........ ........ 19% 310M 1s 55296K ........ ........ ........ ........ ........ ........ 20% 319M 1s 58368K ........ ........ ........ ........ ........ ........ 21% 310M 1s 61440K ........ ........ ........ ........ ........ ........ 22% 235M 1s 64512K ........ ........ ........ ........ ........ ........ 23% 274M 1s 67584K ........ ........ ........ ........ ........ ........ 24% 336M 1s 70656K ........ ........ ........ ........ ........ ........ 25% 222M 1s 73728K ........ ........ ........ ........ ........ ........ 27% 307M 1s 76800K ........ ........ ........ ........ ........ ........ 28% 328M 1s 79872K ........ ........ ........ ........ ........ ........ 29% 227M 1s 82944K ........ ........ ........ ........ ........ ........ 30% 301M 1s 86016K ........ ........ ........ ........ ........ ........ 31% 326M 1s 89088K ........ ........ ........ ........ ........ ........ 32% 273M 1s 92160K ........ ........ ........ ........ ........ ........ 33% 269M 1s 95232K ........ ........ ........ ........ ........ ........ 34% 281M 1s 98304K ........ ........ ........ ........ ........ ........ 35% 255M 1s 101376K ........ ........ ........ ........ ........ ........ 36% 282M 1s 104448K ........ ........ ........ ........ ........ ........ 37% 346M 1s 107520K ........ ........ ........ ........ ........ ........ 38% 305M 1s 110592K ........ ........ ........ ........ ........ ........ 40% 259M 1s 113664K ........ ........ ........ ........ ........ ........ 41% 269M 1s 116736K ........ ........ ........ ........ ........ ........ 42% 307M 1s 119808K ........ ........ ........ ........ ........ ........ 43% 319M 1s 122880K ........ ........ ........ ........ ........ ........ 44% 329M 1s 125952K ........ ........ ........ ........ ........ ........ 45% 284M 1s 129024K ........ ........ ........ ........ ........ ........ 46% 179M 1s 132096K ........ ........ ........ ........ ........ ........ 47% 287M 1s 135168K ........ ........ ........ ........ ........ ........ 48% 331M 1s 138240K ........ ........ ........ ........ ........ ........ 49% 233M 1s 141312K ........ ........ ........ ........ ........ ........ 50% 296M 1s 144384K ........ ........ ........ ........ ........ ........ 51% 281M 1s 147456K ........ ........ ........ ........ ........ ........ 53% 260M 1s 150528K ........ ........ ........ ........ ........ ........ 54% 268M 1s 153600K ........ ........ ........ ........ ........ ........ 55% 370M 1s 156672K ........ ........ ........ ........ ........ ........ 56% 294M 1s 159744K ........ ........ ........ ........ ........ ........ 57% 312M 0s 162816K ........ ........ ........ ........ ........ ........ 58% 328M 0s 165888K ........ ........ ........ ........ ........ ........ 59% 329M 0s 168960K ........ ........ ........ ........ ........ ........ 60% 326M 0s 172032K ........ ........ ........ ........ ........ ........ 61% 323M 0s 175104K ........ ........ ........ ........ ........ ........ 62% 325M 0s 178176K ........ ........ ........ ........ ........ ........ 63% 338M 0s 181248K ........ ........ ........ ........ ........ ........ 64% 332M 0s 184320K ........ ........ ........ ........ ........ ........ 66% 237M 0s 187392K ........ ........ ........ ........ ........ ........ 67% 252M 0s 190464K ........ ........ ........ ........ ........ ........ 68% 295M 0s 193536K ........ ........ ........ ........ ........ ........ 69% 279M 0s 196608K ........ ........ ........ ........ ........ ........ 70% 272M 0s 199680K ........ ........ ........ ........ ........ ........ 71% 259M 0s 202752K ........ ........ ........ ........ ........ ........ 72% 274M 0s 205824K ........ ........ ........ ........ ........ ........ 73% 262M 0s 208896K ........ ........ ........ ........ ........ ........ 74% 248M 0s 211968K ........ ........ ........ ........ ........ ........ 75% 258M 0s 215040K ........ ........ ........ ........ ........ ........ 76% 286M 0s 218112K ........ ........ ........ ........ ........ ........ 77% 361M 0s 221184K ........ ........ ........ ........ ........ ........ 79% 336M 0s 224256K ........ ........ ........ ........ ........ ........ 80% 344M 0s 227328K ........ ........ ........ ........ ........ ........ 81% 222M 0s 230400K ........ ........ ........ ........ ........ ........ 82% 280M 0s 233472K ........ ........ ........ ........ ........ ........ 83% 258M 0s 236544K ........ ........ ........ ........ ........ ........ 84% 212M 0s 239616K ........ ........ ........ ........ ........ ........ 85% 296M 0s 242688K ........ ........ ........ ........ ........ ........ 86% 303M 0s 245760K ........ ........ ........ ........ ........ ........ 87% 185M 0s 248832K ........ ........ ........ ........ ........ ........ 88% 293M 0s 251904K ........ ........ ........ ........ ........ ........ 89% 292M 0s 254976K ........ ........ ........ ........ ........ ........ 90% 324M 0s 258048K ........ ........ ........ ........ ........ ........ 92% 334M 0s 261120K ........ ........ ........ ........ ........ ........ 93% 338M 0s 264192K ........ ........ ........ ........ ........ ........ 94% 208M 0s 267264K ........ ........ ........ ........ ........ ........ 95% 279M 0s 270336K ........ ........ ........ ........ ........ ........ 96% 331M 0s 273408K ........ ........ ........ ........ ........ ........ 97% 338M 0s 276480K ........ ........ ........ ........ ........ ........ 98% 328M 0s 279552K ........ ........ ........ ........ ........ ........ 99% 309M 0s 282624K ........ ........ 100% 357M=1.1s 2024-04-22 04:14:28 (259 MB/s) - ‘karaf-0.21.0.zip’ saved [290466339/290466339] Extracting the new controller... + echo 'Extracting the new controller...' + unzip -q karaf-0.21.0.zip Adding external repositories... + echo 'Adding external repositories...' + sed -ie 's%org.ops4j.pax.url.mvn.repositories=%org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases%g' /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg + cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # # If set to true, the following property will not allow any certificate to be used # when accessing Maven repositories through SSL # #org.ops4j.pax.url.mvn.certificateCheck= # # Path to the local Maven settings file. # The repositories defined in this file will be automatically added to the list # of default repositories if the 'org.ops4j.pax.url.mvn.repositories' property # below is not set. # The following locations are checked for the existence of the settings.xml file # * 1. looks for the specified url # * 2. if not found looks for ${user.home}/.m2/settings.xml # * 3. if not found looks for ${maven.home}/conf/settings.xml # * 4. if not found looks for ${M2_HOME}/conf/settings.xml # #org.ops4j.pax.url.mvn.settings= # # Path to the local Maven repository which is used to avoid downloading # artifacts when they already exist locally. # The value of this property will be extracted from the settings.xml file # above, or defaulted to: # System.getProperty( "user.home" ) + "/.m2/repository" # org.ops4j.pax.url.mvn.localRepository=${karaf.home}/${karaf.default.repository} # # Default this to false. It's just weird to use undocumented repos # org.ops4j.pax.url.mvn.useFallbackRepositories=false # # Uncomment if you don't wanna use the proxy settings # from the Maven conf/settings.xml file # # org.ops4j.pax.url.mvn.proxySupport=false # # Comma separated list of repositories scanned when resolving an artifact. # Those repositories will be checked before iterating through the # below list of repositories and even before the local repository # A repository url can be appended with zero or more of the following flags: # @snapshots : the repository contains snaphots # @noreleases : the repository does not contain any released artifacts # # The following property value will add the system folder as a repo. # org.ops4j.pax.url.mvn.defaultRepositories=\ file:${karaf.home}/${karaf.default.repository}@id=system.repository@snapshots,\ file:${karaf.data}/kar@id=kar.repository@multi@snapshots,\ file:${karaf.base}/${karaf.default.repository}@id=child.system.repository@snapshots # Use the default local repo (e.g.~/.m2/repository) as a "remote" repo #org.ops4j.pax.url.mvn.defaultLocalRepoAsRemote=false # # Comma separated list of repositories scanned when resolving an artifact. # The default list includes the following repositories: # http://repo1.maven.org/maven2@id=central # http://repository.springsource.com/maven/bundles/release@id=spring.ebr # http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external # http://zodiac.springsource.com/maven/bundles/release@id=gemini # http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases # https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases # https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases # To add repositories to the default ones, prepend '+' to the list of repositories # to add. # A repository url can be appended with zero or more of the following flags: # @snapshots : the repository contains snapshots # @noreleases : the repository does not contain any released artifacts # @id=repository.id : the id for the repository, just like in the settings.xml this is optional but recommended # org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases ### ^^^ No remote repositories. This is the only ODL change compared to Karaf defaults.Configuring the startup features... + [[ True == \T\r\u\e ]] + echo 'Configuring the startup features...' + sed -ie 's/\(featuresBoot=\|featuresBoot =\)/featuresBoot = odl-infrautils-ready,odl-jolokia,odl-openflowplugin-flow-services-rest,odl-openflowplugin-app-table-miss-enforcer,/g' /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg + FEATURE_TEST_STRING=features-test + FEATURE_TEST_VERSION=0.21.0 + KARAF_VERSION=karaf4 + [[ integration == \i\n\t\e\g\r\a\t\i\o\n ]] + sed -ie 's%\(featuresRepositories=\|featuresRepositories =\)%featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.0/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features,%g' /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg + [[ ! -z '' ]] + cat /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # # Comma separated list of features repositories to register by default # featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.0/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features, file:${karaf.etc}/26fa34cc-b631-49df-84f3-36941131d37f.xml # # Comma separated list of features to install at startup # featuresBoot = odl-infrautils-ready,odl-jolokia,odl-openflowplugin-flow-services-rest,odl-openflowplugin-app-table-miss-enforcer, e5317303-342d-4134-9449-c8c22a0bc5b4 # # Resource repositories (OBR) that the features resolver can use # to resolve requirements/capabilities # # The format of the resourceRepositories is # resourceRepositories=[xml:url|json:url],... # for Instance: # #resourceRepositories=xml:http://host/path/to/index.xml # or #resourceRepositories=json:http://host/path/to/index.json # # # Defines if the boot features are started in asynchronous mode (in a dedicated thread) # featuresBootAsynchronous=false # # Service requirements enforcement # # By default, the feature resolver checks the service requirements/capabilities of # bundles for new features (xml schema >= 1.3.0) in order to automatically installs # the required bundles. # The following flag can have those values: # - disable: service requirements are completely ignored # - default: service requirements are ignored for old features # - enforce: service requirements are always verified # #serviceRequirements=default # # Store cfg file for config element in feature # #configCfgStore=true # # Define if the feature service automatically refresh bundles # autoRefresh=true # # Configuration of features processing mechanism (overrides, blacklisting, modification of features) # XML file defines instructions related to features processing # versions.properties may declare properties to resolve placeholders in XML file # both files are relative to ${karaf.etc} # #featureProcessing=org.apache.karaf.features.xml #featureProcessingVersions=versions.properties + configure_karaf_log karaf4 '' + local -r karaf_version=karaf4 + local -r controllerdebugmap= + local logapi=log4j + grep log4j2 /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n log4j2.rootLogger.level = INFO #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 64MB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE Configuring the karaf log... karaf_version: karaf4, logapi: log4j2 + logapi=log4j2 + echo 'Configuring the karaf log... karaf_version: karaf4, logapi: log4j2' + '[' log4j2 == log4j2 ']' + sed -ie 's/log4j2.appender.rolling.policies.size.size = 64MB/log4j2.appender.rolling.policies.size.size = 1GB/g' /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg controllerdebugmap: cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg + orgmodule=org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver + orgmodule_=org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver + echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN' + echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN' + unset IFS + echo 'controllerdebugmap: ' + '[' -n '' ']' + echo 'cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg' + cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # Common pattern layout for appenders log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n # Root logger log4j2.rootLogger.level = INFO # uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} # Filters for logs marked by org.opendaylight.odlparent.Markers log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL # Loggers configuration # Spifly logger log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN # Security audit logger log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Appenders configuration # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} # Rolling file appender log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i # uncomment to not force a disk flush #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 1GB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 # Audit file appender log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 # OSGi appender log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * # help with identification of maven-related problems with pax-url-aether #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN Configure java home: /usr/lib/jvm/java-17-openjdk max memory: 2048m memconf: /tmp/karaf-0.21.0/bin/setenv + set_java_vars /usr/lib/jvm/java-17-openjdk 2048m /tmp/karaf-0.21.0/bin/setenv + local -r java_home=/usr/lib/jvm/java-17-openjdk + local -r controllermem=2048m + local -r memconf=/tmp/karaf-0.21.0/bin/setenv + echo Configure + echo ' java home: /usr/lib/jvm/java-17-openjdk' + echo ' max memory: 2048m' + echo ' memconf: /tmp/karaf-0.21.0/bin/setenv' + sed -ie 's%^# export JAVA_HOME%export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-17-openjdk}%g' /tmp/karaf-0.21.0/bin/setenv + sed -ie 's/JAVA_MAX_MEM="2048m"/JAVA_MAX_MEM=2048m/g' /tmp/karaf-0.21.0/bin/setenv cat /tmp/karaf-0.21.0/bin/setenv + echo 'cat /tmp/karaf-0.21.0/bin/setenv' + cat /tmp/karaf-0.21.0/bin/setenv #!/bin/sh # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # # handle specific scripts; the SCRIPT_NAME is exactly the name of the Karaf # script: client, instance, shell, start, status, stop, karaf # # if [ "${KARAF_SCRIPT}" == "SCRIPT_NAME" ]; then # Actions go here... # fi # # general settings which should be applied for all scripts go here; please keep # in mind that it is possible that scripts might be executed more than once, e.g. # in example of the start script where the start script is executed first and the # karaf script afterwards. # # # The following section shows the possible configuration options for the default # karaf scripts # export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-17-openjdk} # Location of Java installation # export JAVA_OPTS # Generic JVM options, for instance, where you can pass the memory configuration # export JAVA_NON_DEBUG_OPTS # Additional non-debug JVM options # export EXTRA_JAVA_OPTS # Additional JVM options # export KARAF_HOME # Karaf home folder # export KARAF_DATA # Karaf data folder # export KARAF_BASE # Karaf base folder # export KARAF_ETC # Karaf etc folder # export KARAF_LOG # Karaf log folder # export KARAF_SYSTEM_OPTS # First citizen Karaf options # export KARAF_OPTS # Additional available Karaf options # export KARAF_DEBUG # Enable debug mode # export KARAF_REDIRECT # Enable/set the std/err redirection when using bin/start # export KARAF_NOROOT # Prevent execution as root if set to true Set Java version + echo 'Set Java version' + sudo /usr/sbin/alternatives --install /usr/bin/java java /usr/lib/jvm/java-17-openjdk/bin/java 1 + sudo /usr/sbin/alternatives --set java /usr/lib/jvm/java-17-openjdk/bin/java JDK default version ... + echo 'JDK default version ...' + java -version openjdk version "17.0.6-ea" 2023-01-17 LTS OpenJDK Runtime Environment (Red_Hat-17.0.6.0.9-0.3.ea.el8) (build 17.0.6-ea+9-LTS) OpenJDK 64-Bit Server VM (Red_Hat-17.0.6.0.9-0.3.ea.el8) (build 17.0.6-ea+9-LTS, mixed mode, sharing) Set JAVA_HOME + echo 'Set JAVA_HOME' + export JAVA_HOME=/usr/lib/jvm/java-17-openjdk + JAVA_HOME=/usr/lib/jvm/java-17-openjdk ++ readlink -e /usr/lib/jvm/java-17-openjdk/bin/java Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java Listing all open ports on controller system... + JAVA_RESOLVED=/usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java + echo 'Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java' + echo 'Listing all open ports on controller system...' + netstat -pnatu (Not all processes could be identified, non-owned process info will not be shown, you would have to be root to see it all.) Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN - tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN - tcp 0 0 10.30.170.65:48842 199.204.45.87:443 TIME_WAIT - tcp 0 164 10.30.170.65:22 10.30.171.144:37754 ESTABLISHED - tcp6 0 0 :::22 :::* LISTEN - tcp6 0 0 :::111 :::* LISTEN - tcp6 0 0 :::5555 :::* LISTEN - udp 0 0 10.30.170.65:68 10.30.170.2:67 ESTABLISHED - udp 0 0 0.0.0.0:111 0.0.0.0:* - udp 0 0 127.0.0.1:323 0.0.0.0:* - udp6 0 0 :::111 :::* - udp6 0 0 ::1:323 :::* - Configuring cluster + '[' -f /tmp/custom_shard_config.txt ']' + echo 'Configuring cluster' + /tmp/karaf-0.21.0/bin/configure_cluster.sh 2 10.30.170.145 10.30.170.65 10.30.170.189 ################################################ ## Configure Cluster ## ################################################ NOTE: Cluster configuration files not found. Copying from /tmp/karaf-0.21.0/system/org/opendaylight/controller/sal-clustering-config/9.0.2 Configuring unique name in akka.conf Configuring hostname in akka.conf Configuring data and rpc seed nodes in akka.conf modules = [ { name = "inventory" namespace = "urn:opendaylight:inventory" shard-strategy = "module" }, { name = "topology" namespace = "urn:TBD:params:xml:ns:yang:network-topology" shard-strategy = "module" }, { name = "toaster" namespace = "http://netconfcentral.org/ns/toaster" shard-strategy = "module" } ] Configuring replication type in module-shards.conf ################################################ ## NOTE: Manually restart controller to ## ## apply configuration. ## ################################################ Dump akka.conf + echo 'Dump akka.conf' + cat /tmp/karaf-0.21.0/configuration/initial/akka.conf odl-cluster-data { akka { remote { artery { enabled = on transport = tcp canonical.hostname = "10.30.170.65" canonical.port = 2550 } } cluster { # Using artery. seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.145:2550", "akka://opendaylight-cluster-data@10.30.170.65:2550", "akka://opendaylight-cluster-data@10.30.170.189:2550"] roles = ["member-2"] # when under load we might trip a false positive on the failure detector # failure-detector { # heartbeat-interval = 4 s # acceptable-heartbeat-pause = 16s # } } persistence { # By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by # modifying the following two properties. The directory location specified may be a relative or absolute path. # The relative path is always relative to KARAF_HOME. # snapshot-store.local.dir = "target/snapshots" # Use lz4 compression for LocalSnapshotStore snapshots snapshot-store.local.use-lz4-compression = false # Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB snapshot-store.local.lz4-blocksize = 256KB } disable-default-actor-system-quarantined-event-handling = "false" } } Dump modules.conf + echo 'Dump modules.conf' + cat /tmp/karaf-0.21.0/configuration/initial/modules.conf modules = [ { name = "inventory" namespace = "urn:opendaylight:inventory" shard-strategy = "module" }, { name = "topology" namespace = "urn:TBD:params:xml:ns:yang:network-topology" shard-strategy = "module" }, { name = "toaster" namespace = "http://netconfcentral.org/ns/toaster" shard-strategy = "module" } ] Dump module-shards.conf + echo 'Dump module-shards.conf' + cat /tmp/karaf-0.21.0/configuration/initial/module-shards.conf module-shards = [ { name = "default" shards = [ { name = "default" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "inventory" shards = [ { name="inventory" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "topology" shards = [ { name="topology" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "toaster" shards = [ { name="toaster" replicas = ["member-1", "member-2", "member-3"] } ] } ] Configuring member-3 with IP address 10.30.170.189 Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + source /tmp/common-functions.sh karaf-0.21.0 common-functions.sh is being sourced ++ [[ /tmp/common-functions.sh == \/\t\m\p\/\c\o\n\f\i\g\u\r\a\t\i\o\n\-\s\c\r\i\p\t\.\s\h ]] ++ echo 'common-functions.sh is being sourced' ++ BUNDLEFOLDER=karaf-0.21.0 ++ export MAVENCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ++ MAVENCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ++ export FEATURESCONF=/tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ++ FEATURESCONF=/tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ++ export CUSTOMPROP=/tmp/karaf-0.21.0/etc/custom.properties ++ CUSTOMPROP=/tmp/karaf-0.21.0/etc/custom.properties ++ export LOGCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ++ LOGCONF=/tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ++ export MEMCONF=/tmp/karaf-0.21.0/bin/setenv ++ MEMCONF=/tmp/karaf-0.21.0/bin/setenv ++ export CONTROLLERMEM= ++ CONTROLLERMEM= ++ export AKKACONF=/tmp/karaf-0.21.0/configuration/initial/akka.conf ++ AKKACONF=/tmp/karaf-0.21.0/configuration/initial/akka.conf ++ export MODULESCONF=/tmp/karaf-0.21.0/configuration/initial/modules.conf ++ MODULESCONF=/tmp/karaf-0.21.0/configuration/initial/modules.conf ++ export MODULESHARDSCONF=/tmp/karaf-0.21.0/configuration/initial/module-shards.conf ++ MODULESHARDSCONF=/tmp/karaf-0.21.0/configuration/initial/module-shards.conf ++ print_common_env ++ cat common-functions environment: MAVENCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ACTUALFEATURES: FEATURESCONF: /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg CUSTOMPROP: /tmp/karaf-0.21.0/etc/custom.properties LOGCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg MEMCONF: /tmp/karaf-0.21.0/bin/setenv CONTROLLERMEM: AKKACONF: /tmp/karaf-0.21.0/configuration/initial/akka.conf MODULESCONF: /tmp/karaf-0.21.0/configuration/initial/modules.conf MODULESHARDSCONF: /tmp/karaf-0.21.0/configuration/initial/module-shards.conf SUITES: ++ SSH='ssh -t -t' ++ extra_services_cntl=' dnsmasq.service httpd.service libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service rabbitmq-server.service ' ++ extra_services_cmp=' libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service ' Changing to /tmp + echo 'Changing to /tmp' + cd /tmp Downloading the distribution from https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip + echo 'Downloading the distribution from https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip' + wget --progress=dot:mega https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip --2024-04-22 04:14:31-- https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip Resolving nexus.opendaylight.org (nexus.opendaylight.org)... 199.204.45.87, 2604:e100:1:0:f816:3eff:fe45:48d6 Connecting to nexus.opendaylight.org (nexus.opendaylight.org)|199.204.45.87|:443... connected. HTTP request sent, awaiting response... 200 OK Length: 290466339 (277M) [application/zip] Saving to: ‘karaf-0.21.0.zip’ 0K ........ ........ ........ ........ ........ ........ 1% 72.3M 4s 3072K ........ ........ ........ ........ ........ ........ 2% 145M 3s 6144K ........ ........ ........ ........ ........ ........ 3% 148M 2s 9216K ........ ........ ........ ........ ........ ........ 4% 174M 2s 12288K ........ ........ ........ ........ ........ ........ 5% 176M 2s 15360K ........ ........ ........ ........ ........ ........ 6% 178M 2s 18432K ........ ........ ........ ........ ........ ........ 7% 199M 2s 21504K ........ ........ ........ ........ ........ ........ 8% 197M 2s 24576K ........ ........ ........ ........ ........ ........ 9% 216M 2s 27648K ........ ........ ........ ........ ........ ........ 10% 245M 2s 30720K ........ ........ ........ ........ ........ ........ 11% 245M 1s 33792K ........ ........ ........ ........ ........ ........ 12% 267M 1s 36864K ........ ........ ........ ........ ........ ........ 14% 274M 1s 39936K ........ ........ ........ ........ ........ ........ 15% 279M 1s 43008K ........ ........ ........ ........ ........ ........ 16% 281M 1s 46080K ........ ........ ........ ........ ........ ........ 17% 277M 1s 49152K ........ ........ ........ ........ ........ ........ 18% 279M 1s 52224K ........ ........ ........ ........ ........ ........ 19% 274M 1s 55296K ........ ........ ........ ........ ........ ........ 20% 273M 1s 58368K ........ ........ ........ ........ ........ ........ 21% 281M 1s 61440K ........ ........ ........ ........ ........ ........ 22% 270M 1s 64512K ........ ........ ........ ........ ........ ........ 23% 265M 1s 67584K ........ ........ ........ ........ ........ ........ 24% 268M 1s 70656K ........ ........ ........ ........ ........ ........ 25% 270M 1s 73728K ........ ........ ........ ........ ........ ........ 27% 270M 1s 76800K ........ ........ ........ ........ ........ ........ 28% 272M 1s 79872K ........ ........ ........ ........ ........ ........ 29% 278M 1s 82944K ........ ........ ........ ........ ........ ........ 30% 272M 1s 86016K ........ ........ ........ ........ ........ ........ 31% 295M 1s 89088K ........ ........ ........ ........ ........ ........ 32% 349M 1s 92160K ........ ........ ........ ........ ........ ........ 33% 331M 1s 95232K ........ ........ ........ ........ ........ ........ 34% 342M 1s 98304K ........ ........ ........ ........ ........ ........ 35% 356M 1s 101376K ........ ........ ........ ........ ........ ........ 36% 359M 1s 104448K ........ ........ ........ ........ ........ ........ 37% 346M 1s 107520K ........ ........ ........ ........ ........ ........ 38% 356M 1s 110592K ........ ........ ........ ........ ........ ........ 40% 350M 1s 113664K ........ ........ ........ ........ ........ ........ 41% 352M 1s 116736K ........ ........ ........ ........ ........ ........ 42% 338M 1s 119808K ........ ........ ........ ........ ........ ........ 43% 346M 1s 122880K ........ ........ ........ ........ ........ ........ 44% 338M 1s 125952K ........ ........ ........ ........ ........ ........ 45% 345M 1s 129024K ........ ........ ........ ........ ........ ........ 46% 314M 1s 132096K ........ ........ ........ ........ ........ ........ 47% 359M 1s 135168K ........ ........ ........ ........ ........ ........ 48% 347M 1s 138240K ........ ........ ........ ........ ........ ........ 49% 333M 1s 141312K ........ ........ ........ ........ ........ ........ 50% 338M 1s 144384K ........ ........ ........ ........ ........ ........ 51% 338M 1s 147456K ........ ........ ........ ........ ........ ........ 53% 343M 1s 150528K ........ ........ ........ ........ ........ ........ 54% 343M 0s 153600K ........ ........ ........ ........ ........ ........ 55% 341M 0s 156672K ........ ........ ........ ........ ........ ........ 56% 319M 0s 159744K ........ ........ ........ ........ ........ ........ 57% 294M 0s 162816K ........ ........ ........ ........ ........ ........ 58% 297M 0s 165888K ........ ........ ........ ........ ........ ........ 59% 285M 0s 168960K ........ ........ ........ ........ ........ ........ 60% 301M 0s 172032K ........ ........ ........ ........ ........ ........ 61% 302M 0s 175104K ........ ........ ........ ........ ........ ........ 62% 271M 0s 178176K ........ ........ ........ ........ ........ ........ 63% 300M 0s 181248K ........ ........ ........ ........ ........ ........ 64% 348M 0s 184320K ........ ........ ........ ........ ........ ........ 66% 338M 0s 187392K ........ ........ ........ ........ ........ ........ 67% 335M 0s 190464K ........ ........ ........ ........ ........ ........ 68% 346M 0s 193536K ........ ........ ........ ........ ........ ........ 69% 192M 0s 196608K ........ ........ ........ ........ ........ ........ 70% 134M 0s 199680K ........ ........ ........ ........ ........ ........ 71% 338M 0s 202752K ........ ........ ........ ........ ........ ........ 72% 150M 0s 205824K ........ ........ ........ ........ ........ ........ 73% 392M 0s 208896K ........ ........ ........ ........ ........ ........ 74% 390M 0s 211968K ........ ........ ........ ........ ........ ........ 75% 392M 0s 215040K ........ ........ ........ ........ ........ ........ 76% 395M 0s 218112K ........ ........ ........ ........ ........ ........ 77% 387M 0s 221184K ........ ........ ........ ........ ........ ........ 79% 287M 0s 224256K ........ ........ ........ ........ ........ ........ 80% 337M 0s 227328K ........ ........ ........ ........ ........ ........ 81% 343M 0s 230400K ........ ........ ........ ........ ........ ........ 82% 347M 0s 233472K ........ ........ ........ ........ ........ ........ 83% 343M 0s 236544K ........ ........ ........ ........ ........ ........ 84% 340M 0s 239616K ........ ........ ........ ........ ........ ........ 85% 374M 0s 242688K ........ ........ ........ ........ ........ ........ 86% 373M 0s 245760K ........ ........ ........ ........ ........ ........ 87% 355M 0s 248832K ........ ........ ........ ........ ........ ........ 88% 378M 0s 251904K ........ ........ ........ ........ ........ ........ 89% 364M 0s 254976K ........ ........ ........ ........ ........ ........ 90% 373M 0s 258048K ........ ........ ........ ........ ........ ........ 92% 365M 0s 261120K ........ ........ ........ ........ ........ ........ 93% 358M 0s 264192K ........ ........ ........ ........ ........ ........ 94% 374M 0s 267264K ........ ........ ........ ........ ........ ........ 95% 369M 0s 270336K ........ ........ ........ ........ ........ ........ 96% 353M 0s 273408K ........ ........ ........ ........ ........ ........ 97% 288M 0s 276480K ........ ........ ........ ........ ........ ........ 98% 274M 0s 279552K ........ ........ ........ ........ ........ ........ 99% 361M 0s 282624K ........ ........ 100% 379M=1.0s 2024-04-22 04:14:32 (280 MB/s) - ‘karaf-0.21.0.zip’ saved [290466339/290466339] Extracting the new controller... + echo 'Extracting the new controller...' + unzip -q karaf-0.21.0.zip Adding external repositories... + echo 'Adding external repositories...' + sed -ie 's%org.ops4j.pax.url.mvn.repositories=%org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases%g' /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg + cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # # If set to true, the following property will not allow any certificate to be used # when accessing Maven repositories through SSL # #org.ops4j.pax.url.mvn.certificateCheck= # # Path to the local Maven settings file. # The repositories defined in this file will be automatically added to the list # of default repositories if the 'org.ops4j.pax.url.mvn.repositories' property # below is not set. # The following locations are checked for the existence of the settings.xml file # * 1. looks for the specified url # * 2. if not found looks for ${user.home}/.m2/settings.xml # * 3. if not found looks for ${maven.home}/conf/settings.xml # * 4. if not found looks for ${M2_HOME}/conf/settings.xml # #org.ops4j.pax.url.mvn.settings= # # Path to the local Maven repository which is used to avoid downloading # artifacts when they already exist locally. # The value of this property will be extracted from the settings.xml file # above, or defaulted to: # System.getProperty( "user.home" ) + "/.m2/repository" # org.ops4j.pax.url.mvn.localRepository=${karaf.home}/${karaf.default.repository} # # Default this to false. It's just weird to use undocumented repos # org.ops4j.pax.url.mvn.useFallbackRepositories=false # # Uncomment if you don't wanna use the proxy settings # from the Maven conf/settings.xml file # # org.ops4j.pax.url.mvn.proxySupport=false # # Comma separated list of repositories scanned when resolving an artifact. # Those repositories will be checked before iterating through the # below list of repositories and even before the local repository # A repository url can be appended with zero or more of the following flags: # @snapshots : the repository contains snaphots # @noreleases : the repository does not contain any released artifacts # # The following property value will add the system folder as a repo. # org.ops4j.pax.url.mvn.defaultRepositories=\ file:${karaf.home}/${karaf.default.repository}@id=system.repository@snapshots,\ file:${karaf.data}/kar@id=kar.repository@multi@snapshots,\ file:${karaf.base}/${karaf.default.repository}@id=child.system.repository@snapshots # Use the default local repo (e.g.~/.m2/repository) as a "remote" repo #org.ops4j.pax.url.mvn.defaultLocalRepoAsRemote=false # # Comma separated list of repositories scanned when resolving an artifact. # The default list includes the following repositories: # http://repo1.maven.org/maven2@id=central # http://repository.springsource.com/maven/bundles/release@id=spring.ebr # http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external # http://zodiac.springsource.com/maven/bundles/release@id=gemini # http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases # https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases # https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases # To add repositories to the default ones, prepend '+' to the list of repositories # to add. # A repository url can be appended with zero or more of the following flags: # @snapshots : the repository contains snapshots # @noreleases : the repository does not contain any released artifacts # @id=repository.id : the id for the repository, just like in the settings.xml this is optional but recommended # org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases ### ^^^ No remote repositories. This is the only ODL change compared to Karaf defaults.Configuring the startup features... + [[ True == \T\r\u\e ]] + echo 'Configuring the startup features...' + sed -ie 's/\(featuresBoot=\|featuresBoot =\)/featuresBoot = odl-infrautils-ready,odl-jolokia,odl-openflowplugin-flow-services-rest,odl-openflowplugin-app-table-miss-enforcer,/g' /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg + FEATURE_TEST_STRING=features-test + FEATURE_TEST_VERSION=0.21.0 + KARAF_VERSION=karaf4 + [[ integration == \i\n\t\e\g\r\a\t\i\o\n ]] + sed -ie 's%\(featuresRepositories=\|featuresRepositories =\)%featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.0/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features,%g' /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg + [[ ! -z '' ]] + cat /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # # Comma separated list of features repositories to register by default # featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.0/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features, file:${karaf.etc}/26fa34cc-b631-49df-84f3-36941131d37f.xml # # Comma separated list of features to install at startup # featuresBoot = odl-infrautils-ready,odl-jolokia,odl-openflowplugin-flow-services-rest,odl-openflowplugin-app-table-miss-enforcer, e5317303-342d-4134-9449-c8c22a0bc5b4 # # Resource repositories (OBR) that the features resolver can use # to resolve requirements/capabilities # # The format of the resourceRepositories is # resourceRepositories=[xml:url|json:url],... # for Instance: # #resourceRepositories=xml:http://host/path/to/index.xml # or #resourceRepositories=json:http://host/path/to/index.json # # # Defines if the boot features are started in asynchronous mode (in a dedicated thread) # featuresBootAsynchronous=false # # Service requirements enforcement # # By default, the feature resolver checks the service requirements/capabilities of # bundles for new features (xml schema >= 1.3.0) in order to automatically installs # the required bundles. # The following flag can have those values: # - disable: service requirements are completely ignored # - default: service requirements are ignored for old features # - enforce: service requirements are always verified # #serviceRequirements=default # # Store cfg file for config element in feature # #configCfgStore=true # # Define if the feature service automatically refresh bundles # autoRefresh=true # # Configuration of features processing mechanism (overrides, blacklisting, modification of features) # XML file defines instructions related to features processing # versions.properties may declare properties to resolve placeholders in XML file # both files are relative to ${karaf.etc} # #featureProcessing=org.apache.karaf.features.xml #featureProcessingVersions=versions.properties + configure_karaf_log karaf4 '' + local -r karaf_version=karaf4 + local -r controllerdebugmap= + local logapi=log4j + grep log4j2 /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n log4j2.rootLogger.level = INFO #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 64MB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE + logapi=log4j2 + echo 'Configuring the karaf log... karaf_version: karaf4, logapi: log4j2' Configuring the karaf log... karaf_version: karaf4, logapi: log4j2 + '[' log4j2 == log4j2 ']' + sed -ie 's/log4j2.appender.rolling.policies.size.size = 64MB/log4j2.appender.rolling.policies.size.size = 1GB/g' /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg + orgmodule=org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver + orgmodule_=org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver + echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN' controllerdebugmap: cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg + echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN' + unset IFS + echo 'controllerdebugmap: ' + '[' -n '' ']' + echo 'cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg' + cat /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # Common pattern layout for appenders log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n # Root logger log4j2.rootLogger.level = INFO # uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} # Filters for logs marked by org.opendaylight.odlparent.Markers log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL # Loggers configuration # Spifly logger log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN # Security audit logger log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Appenders configuration # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} # Rolling file appender log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i # uncomment to not force a disk flush #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 1GB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 # Audit file appender log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 # OSGi appender log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * # help with identification of maven-related problems with pax-url-aether #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN + set_java_vars /usr/lib/jvm/java-17-openjdk 2048m /tmp/karaf-0.21.0/bin/setenv Configure java home: /usr/lib/jvm/java-17-openjdk max memory: 2048m memconf: /tmp/karaf-0.21.0/bin/setenv + local -r java_home=/usr/lib/jvm/java-17-openjdk + local -r controllermem=2048m + local -r memconf=/tmp/karaf-0.21.0/bin/setenv + echo Configure + echo ' java home: /usr/lib/jvm/java-17-openjdk' + echo ' max memory: 2048m' + echo ' memconf: /tmp/karaf-0.21.0/bin/setenv' + sed -ie 's%^# export JAVA_HOME%export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-17-openjdk}%g' /tmp/karaf-0.21.0/bin/setenv + sed -ie 's/JAVA_MAX_MEM="2048m"/JAVA_MAX_MEM=2048m/g' /tmp/karaf-0.21.0/bin/setenv cat /tmp/karaf-0.21.0/bin/setenv + echo 'cat /tmp/karaf-0.21.0/bin/setenv' + cat /tmp/karaf-0.21.0/bin/setenv #!/bin/sh # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # # handle specific scripts; the SCRIPT_NAME is exactly the name of the Karaf # script: client, instance, shell, start, status, stop, karaf # # if [ "${KARAF_SCRIPT}" == "SCRIPT_NAME" ]; then # Actions go here... # fi # # general settings which should be applied for all scripts go here; please keep # in mind that it is possible that scripts might be executed more than once, e.g. # in example of the start script where the start script is executed first and the # karaf script afterwards. # # # The following section shows the possible configuration options for the default # karaf scripts # export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-17-openjdk} # Location of Java installation # export JAVA_OPTS # Generic JVM options, for instance, where you can pass the memory configuration # export JAVA_NON_DEBUG_OPTS # Additional non-debug JVM options # export EXTRA_JAVA_OPTS # Additional JVM options # export KARAF_HOME # Karaf home folder # export KARAF_DATA # Karaf data folder # export KARAF_BASE # Karaf base folder # export KARAF_ETC # Karaf etc folder # export KARAF_LOG # Karaf log folder # export KARAF_SYSTEM_OPTS # First citizen Karaf options # export KARAF_OPTS # Additional available Karaf options # export KARAF_DEBUG # Enable debug mode # export KARAF_REDIRECT # Enable/set the std/err redirection when using bin/start # export KARAF_NOROOT # Prevent execution as root if set to true Set Java version + echo 'Set Java version' + sudo /usr/sbin/alternatives --install /usr/bin/java java /usr/lib/jvm/java-17-openjdk/bin/java 1 + sudo /usr/sbin/alternatives --set java /usr/lib/jvm/java-17-openjdk/bin/java JDK default version ... + echo 'JDK default version ...' + java -version openjdk version "17.0.6-ea" 2023-01-17 LTS OpenJDK Runtime Environment (Red_Hat-17.0.6.0.9-0.3.ea.el8) (build 17.0.6-ea+9-LTS) OpenJDK 64-Bit Server VM (Red_Hat-17.0.6.0.9-0.3.ea.el8) (build 17.0.6-ea+9-LTS, mixed mode, sharing) Set JAVA_HOME + echo 'Set JAVA_HOME' + export JAVA_HOME=/usr/lib/jvm/java-17-openjdk + JAVA_HOME=/usr/lib/jvm/java-17-openjdk ++ readlink -e /usr/lib/jvm/java-17-openjdk/bin/java Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java + JAVA_RESOLVED=/usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java + echo 'Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-17-openjdk-17.0.6.0.9-0.3.ea.el8.x86_64/bin/java' Listing all open ports on controller system... + echo 'Listing all open ports on controller system...' + netstat -pnatu (Not all processes could be identified, non-owned process info will not be shown, you would have to be root to see it all.) Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN - tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN - tcp 0 164 10.30.170.189:22 10.30.171.144:57316 ESTABLISHED - tcp 0 0 10.30.170.189:54372 199.204.45.87:443 TIME_WAIT - tcp6 0 0 :::111 :::* LISTEN - tcp6 0 0 :::5555 :::* LISTEN - tcp6 0 0 :::22 :::* LISTEN - udp 0 0 127.0.0.1:323 0.0.0.0:* - udp 0 0 10.30.170.189:68 10.30.170.2:67 ESTABLISHED - udp 0 0 0.0.0.0:111 0.0.0.0:* - udp6 0 0 ::1:323 :::* - udp6 0 0 :::111 :::* - Configuring cluster + '[' -f /tmp/custom_shard_config.txt ']' + echo 'Configuring cluster' + /tmp/karaf-0.21.0/bin/configure_cluster.sh 3 10.30.170.145 10.30.170.65 10.30.170.189 ################################################ ## Configure Cluster ## ################################################ NOTE: Cluster configuration files not found. Copying from /tmp/karaf-0.21.0/system/org/opendaylight/controller/sal-clustering-config/9.0.2 Configuring unique name in akka.conf Configuring hostname in akka.conf Configuring data and rpc seed nodes in akka.conf modules = [ { name = "inventory" namespace = "urn:opendaylight:inventory" shard-strategy = "module" }, { name = "topology" namespace = "urn:TBD:params:xml:ns:yang:network-topology" shard-strategy = "module" }, { name = "toaster" namespace = "http://netconfcentral.org/ns/toaster" shard-strategy = "module" } ] Configuring replication type in module-shards.conf ################################################ ## NOTE: Manually restart controller to ## ## apply configuration. ## ################################################ Dump akka.conf + echo 'Dump akka.conf' + cat /tmp/karaf-0.21.0/configuration/initial/akka.conf odl-cluster-data { akka { remote { artery { enabled = on transport = tcp canonical.hostname = "10.30.170.189" canonical.port = 2550 } } cluster { # Using artery. seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.145:2550", "akka://opendaylight-cluster-data@10.30.170.65:2550", "akka://opendaylight-cluster-data@10.30.170.189:2550"] roles = ["member-3"] # when under load we might trip a false positive on the failure detector # failure-detector { # heartbeat-interval = 4 s # acceptable-heartbeat-pause = 16s # } } persistence { # By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by # modifying the following two properties. The directory location specified may be a relative or absolute path. # The relative path is always relative to KARAF_HOME. # snapshot-store.local.dir = "target/snapshots" # Use lz4 compression for LocalSnapshotStore snapshots snapshot-store.local.use-lz4-compression = false # Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB snapshot-store.local.lz4-blocksize = 256KB } disable-default-actor-system-quarantined-event-handling = "false" } } + echo 'Dump modules.conf' Dump modules.conf + cat /tmp/karaf-0.21.0/configuration/initial/modules.conf modules = [ { name = "inventory" namespace = "urn:opendaylight:inventory" shard-strategy = "module" }, { name = "topology" namespace = "urn:TBD:params:xml:ns:yang:network-topology" shard-strategy = "module" }, { name = "toaster" namespace = "http://netconfcentral.org/ns/toaster" shard-strategy = "module" } ] Dump module-shards.conf + echo 'Dump module-shards.conf' + cat /tmp/karaf-0.21.0/configuration/initial/module-shards.conf module-shards = [ { name = "default" shards = [ { name = "default" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "inventory" shards = [ { name="inventory" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "topology" shards = [ { name="topology" replicas = ["member-1", "member-2", "member-3"] } ] }, { name = "toaster" shards = [ { name="toaster" replicas = ["member-1", "member-2", "member-3"] } ] } ] Locating config plan to use... config plan exists!!! Changing the config plan path... # Place the suites in run order: /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/scripts/set_akka_debug.sh Executing /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/scripts/set_akka_debug.sh... Copying config files to ODL Controller folder Set AKKA debug on 10.30.170.145 Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. Enable AKKA debug Dump /tmp/karaf-0.21.0/configuration/initial/akka.conf odl-cluster-data { akka { loglevel = "DEBUG" actor { debug { autoreceive = on lifecycle = on unhandled = on fsm = on event-stream = on } } remote { artery { enabled = on transport = tcp canonical.hostname = "10.30.170.145" canonical.port = 2550 } } cluster { # Using artery. seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.145:2550", "akka://opendaylight-cluster-data@10.30.170.65:2550", "akka://opendaylight-cluster-data@10.30.170.189:2550"] roles = ["member-1"] # when under load we might trip a false positive on the failure detector # failure-detector { # heartbeat-interval = 4 s # acceptable-heartbeat-pause = 16s # } } persistence { # By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by # modifying the following two properties. The directory location specified may be a relative or absolute path. # The relative path is always relative to KARAF_HOME. # snapshot-store.local.dir = "target/snapshots" # Use lz4 compression for LocalSnapshotStore snapshots snapshot-store.local.use-lz4-compression = false # Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB snapshot-store.local.lz4-blocksize = 256KB } disable-default-actor-system-quarantined-event-handling = "false" } } Dump /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # Common pattern layout for appenders log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n # Root logger log4j2.rootLogger.level = INFO # uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} # Filters for logs marked by org.opendaylight.odlparent.Markers log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL # Loggers configuration # Spifly logger log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN # Security audit logger log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Appenders configuration # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} # Rolling file appender log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i # uncomment to not force a disk flush #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 1GB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 # Audit file appender log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 # OSGi appender log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * # help with identification of maven-related problems with pax-url-aether #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN log4j2.logger.cluster.name=akka.cluster log4j2.logger.cluster.level=DEBUG log4j2.logger.remote.name=akka.remote log4j2.logger.remote.level=DEBUG Set AKKA debug on 10.30.170.65 Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. Enable AKKA debug Dump /tmp/karaf-0.21.0/configuration/initial/akka.conf odl-cluster-data { akka { loglevel = "DEBUG" actor { debug { autoreceive = on lifecycle = on unhandled = on fsm = on event-stream = on } } remote { artery { enabled = on transport = tcp canonical.hostname = "10.30.170.65" canonical.port = 2550 } } cluster { # Using artery. seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.145:2550", "akka://opendaylight-cluster-data@10.30.170.65:2550", "akka://opendaylight-cluster-data@10.30.170.189:2550"] roles = ["member-2"] # when under load we might trip a false positive on the failure detector # failure-detector { # heartbeat-interval = 4 s # acceptable-heartbeat-pause = 16s # } } persistence { # By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by # modifying the following two properties. The directory location specified may be a relative or absolute path. # The relative path is always relative to KARAF_HOME. # snapshot-store.local.dir = "target/snapshots" # Use lz4 compression for LocalSnapshotStore snapshots snapshot-store.local.use-lz4-compression = false # Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB snapshot-store.local.lz4-blocksize = 256KB } disable-default-actor-system-quarantined-event-handling = "false" } } Dump /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # Common pattern layout for appenders log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n # Root logger log4j2.rootLogger.level = INFO # uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} # Filters for logs marked by org.opendaylight.odlparent.Markers log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL # Loggers configuration # Spifly logger log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN # Security audit logger log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Appenders configuration # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} # Rolling file appender log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i # uncomment to not force a disk flush #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 1GB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 # Audit file appender log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 # OSGi appender log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * # help with identification of maven-related problems with pax-url-aether #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN log4j2.logger.cluster.name=akka.cluster log4j2.logger.cluster.level=DEBUG log4j2.logger.remote.name=akka.remote log4j2.logger.remote.level=DEBUG Set AKKA debug on 10.30.170.189 Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. Enable AKKA debug Dump /tmp/karaf-0.21.0/configuration/initial/akka.conf odl-cluster-data { akka { loglevel = "DEBUG" actor { debug { autoreceive = on lifecycle = on unhandled = on fsm = on event-stream = on } } remote { artery { enabled = on transport = tcp canonical.hostname = "10.30.170.189" canonical.port = 2550 } } cluster { # Using artery. seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.145:2550", "akka://opendaylight-cluster-data@10.30.170.65:2550", "akka://opendaylight-cluster-data@10.30.170.189:2550"] roles = ["member-3"] # when under load we might trip a false positive on the failure detector # failure-detector { # heartbeat-interval = 4 s # acceptable-heartbeat-pause = 16s # } } persistence { # By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by # modifying the following two properties. The directory location specified may be a relative or absolute path. # The relative path is always relative to KARAF_HOME. # snapshot-store.local.dir = "target/snapshots" # Use lz4 compression for LocalSnapshotStore snapshots snapshot-store.local.use-lz4-compression = false # Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB snapshot-store.local.lz4-blocksize = 256KB } disable-default-actor-system-quarantined-event-handling = "false" } } Dump /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg ################################################################################ # # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ################################################################################ # Common pattern layout for appenders log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n # Root logger log4j2.rootLogger.level = INFO # uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library #log4j2.rootLogger.type = asyncRoot #log4j2.rootLogger.includeLocation = false log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi log4j2.rootLogger.appenderRef.Console.ref = Console log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF} # Filters for logs marked by org.opendaylight.odlparent.Markers log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL # Loggers configuration # Spifly logger log4j2.logger.spifly.name = org.apache.aries.spifly log4j2.logger.spifly.level = WARN # Security audit logger log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit log4j2.logger.audit.level = INFO log4j2.logger.audit.additivity = false log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile # Appenders configuration # Console appender not used by default (see log4j2.rootLogger.appenderRefs) log4j2.appender.console.type = Console log4j2.appender.console.name = Console log4j2.appender.console.layout.type = PatternLayout log4j2.appender.console.layout.pattern = ${log4j2.pattern} # Rolling file appender log4j2.appender.rolling.type = RollingRandomAccessFile log4j2.appender.rolling.name = RollingFile log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i # uncomment to not force a disk flush #log4j2.appender.rolling.immediateFlush = false log4j2.appender.rolling.append = true log4j2.appender.rolling.layout.type = PatternLayout log4j2.appender.rolling.layout.pattern = ${log4j2.pattern} log4j2.appender.rolling.policies.type = Policies log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.rolling.policies.size.size = 1GB log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy log4j2.appender.rolling.strategy.max = 7 # Audit file appender log4j2.appender.audit.type = RollingRandomAccessFile log4j2.appender.audit.name = AuditRollingFile log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i log4j2.appender.audit.append = true log4j2.appender.audit.layout.type = PatternLayout log4j2.appender.audit.layout.pattern = ${log4j2.pattern} log4j2.appender.audit.policies.type = Policies log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy log4j2.appender.audit.policies.size.size = 8MB log4j2.appender.audit.strategy.type = DefaultRolloverStrategy log4j2.appender.audit.strategy.max = 7 # OSGi appender log4j2.appender.osgi.type = PaxOsgi log4j2.appender.osgi.name = PaxOsgi log4j2.appender.osgi.filter = * # help with identification of maven-related problems with pax-url-aether #log4j2.logger.aether.name = shaded.org.eclipse.aether #log4j2.logger.aether.level = TRACE #log4j2.logger.http-headers.name = shaded.org.apache.http.headers #log4j2.logger.http-headers.level = DEBUG #log4j2.logger.maven.name = org.ops4j.pax.url.mvn #log4j2.logger.maven.level = TRACE log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN log4j2.logger.cluster.name=akka.cluster log4j2.logger.cluster.level=DEBUG log4j2.logger.remote.name=akka.remote log4j2.logger.remote.level=DEBUG Finished running config plans Starting member-1 with IP address 10.30.170.145 Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. Redirecting karaf console output to karaf_console.log Starting controller... start: Redirecting Karaf output to /tmp/karaf-0.21.0/data/log/karaf_console.log Starting member-2 with IP address 10.30.170.65 Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. Redirecting karaf console output to karaf_console.log Starting controller... start: Redirecting Karaf output to /tmp/karaf-0.21.0/data/log/karaf_console.log Starting member-3 with IP address 10.30.170.189 Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. Redirecting karaf console output to karaf_console.log Starting controller... start: Redirecting Karaf output to /tmp/karaf-0.21.0/data/log/karaf_console.log [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins7559416710283723260.sh common-functions.sh is being sourced common-functions environment: MAVENCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.url.mvn.cfg ACTUALFEATURES: FEATURESCONF: /tmp/karaf-0.21.0/etc/org.apache.karaf.features.cfg CUSTOMPROP: /tmp/karaf-0.21.0/etc/custom.properties LOGCONF: /tmp/karaf-0.21.0/etc/org.ops4j.pax.logging.cfg MEMCONF: /tmp/karaf-0.21.0/bin/setenv CONTROLLERMEM: 2048m AKKACONF: /tmp/karaf-0.21.0/configuration/initial/akka.conf MODULESCONF: /tmp/karaf-0.21.0/configuration/initial/modules.conf MODULESHARDSCONF: /tmp/karaf-0.21.0/configuration/initial/module-shards.conf SUITES: + echo '#################################################' ################################################# + echo '## Verify Cluster is UP ##' ## Verify Cluster is UP ## + echo '#################################################' ################################################# + create_post_startup_script + cat + copy_and_run_post_startup_script + seed_index=1 ++ seq 1 3 + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_1_IP + echo 'Execute the post startup script on controller 10.30.170.145' Execute the post startup script on controller 10.30.170.145 + scp /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/post-startup-script.sh 10.30.170.145:/tmp/ Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + ssh 10.30.170.145 'bash /tmp/post-startup-script.sh 1' Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. tcp6 0 0 :::8101 :::* LISTEN Waiting up to 3 minutes for controller to come up, checking every 5 seconds... 2024-04-22T04:15:15,936 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 205 - org.opendaylight.infrautils.ready-api - 6.0.6 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.] Controller is UP 2024-04-22T04:15:15,936 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 205 - org.opendaylight.infrautils.ready-api - 6.0.6 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.] Listing all open ports on controller system... (Not all processes could be identified, non-owned process info will not be shown, you would have to be root to see it all.) Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN - tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN - tcp 0 248 10.30.170.145:22 10.30.171.144:59662 ESTABLISHED - tcp 0 0 10.30.170.145:54574 199.204.45.87:443 TIME_WAIT - tcp6 0 0 127.0.0.1:44444 :::* LISTEN 5753/java tcp6 0 0 :::6653 :::* LISTEN 5753/java tcp6 0 0 127.0.0.1:33185 :::* LISTEN 5753/java tcp6 0 0 :::8101 :::* LISTEN 5753/java tcp6 0 0 :::6633 :::* LISTEN 5753/java tcp6 0 0 127.0.0.1:1099 :::* LISTEN 5753/java tcp6 0 0 :::111 :::* LISTEN - tcp6 0 0 :::5555 :::* LISTEN - tcp6 0 0 :::8181 :::* LISTEN 5753/java tcp6 0 0 10.30.170.145:2550 :::* LISTEN 5753/java tcp6 0 0 :::22 :::* LISTEN - tcp6 0 0 10.30.170.145:2550 10.30.170.189:42592 ESTABLISHED 5753/java tcp6 0 0 10.30.170.145:2550 10.30.170.189:42588 ESTABLISHED 5753/java tcp6 0 0 10.30.170.145:35788 10.30.170.189:2550 ESTABLISHED 5753/java tcp6 0 0 10.30.170.145:2550 10.30.170.65:52860 ESTABLISHED 5753/java tcp6 0 0 10.30.170.145:2550 10.30.170.65:52870 ESTABLISHED 5753/java tcp6 0 0 127.0.0.1:56826 127.0.0.1:44444 TIME_WAIT - tcp6 0 0 10.30.170.145:37652 10.30.170.65:2550 ESTABLISHED 5753/java tcp6 0 0 10.30.170.145:37646 10.30.170.65:2550 ESTABLISHED 5753/java tcp6 32 0 10.30.170.145:33396 199.204.45.87:443 CLOSE_WAIT 5753/java tcp6 0 0 127.0.0.1:56828 127.0.0.1:44444 TIME_WAIT - tcp6 0 0 10.30.170.145:35782 10.30.170.189:2550 ESTABLISHED 5753/java tcp6 0 0 127.0.0.1:43352 127.0.0.1:1099 TIME_WAIT - udp 0 0 10.30.170.145:68 10.30.170.4:67 ESTABLISHED - udp 0 0 0.0.0.0:111 0.0.0.0:* - udp 0 0 127.0.0.1:323 0.0.0.0:* - udp6 0 0 :::111 :::* - udp6 0 0 ::1:323 :::* - looking for "BindException: Address already in use" in log file looking for "server is unhealthy" in log file + '[' 1 == 0 ']' + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_2_IP + echo 'Execute the post startup script on controller 10.30.170.65' Execute the post startup script on controller 10.30.170.65 + scp /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/post-startup-script.sh 10.30.170.65:/tmp/ Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + ssh 10.30.170.65 'bash /tmp/post-startup-script.sh 2' Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. tcp6 0 0 :::8101 :::* LISTEN Waiting up to 3 minutes for controller to come up, checking every 5 seconds... 2024-04-22T04:15:16,008 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 205 - org.opendaylight.infrautils.ready-api - 6.0.6 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.] Controller is UP 2024-04-22T04:15:16,008 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 205 - org.opendaylight.infrautils.ready-api - 6.0.6 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.] Listing all open ports on controller system... (Not all processes could be identified, non-owned process info will not be shown, you would have to be root to see it all.) Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN - tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN - tcp 0 0 10.30.170.65:48842 199.204.45.87:443 TIME_WAIT - tcp 0 0 10.30.170.65:22 10.30.171.144:56446 ESTABLISHED - tcp6 0 0 :::8181 :::* LISTEN 5751/java tcp6 0 0 10.30.170.65:2550 :::* LISTEN 5751/java tcp6 0 0 :::22 :::* LISTEN - tcp6 0 0 127.0.0.1:44444 :::* LISTEN 5751/java tcp6 0 0 :::6653 :::* LISTEN 5751/java tcp6 0 0 :::8101 :::* LISTEN 5751/java tcp6 0 0 127.0.0.1:33575 :::* LISTEN 5751/java tcp6 0 0 :::6633 :::* LISTEN 5751/java tcp6 0 0 127.0.0.1:1099 :::* LISTEN 5751/java tcp6 0 0 :::111 :::* LISTEN - tcp6 0 0 :::5555 :::* LISTEN - tcp6 0 0 127.0.0.1:57616 127.0.0.1:44444 TIME_WAIT - tcp6 32 0 10.30.170.65:47820 199.204.45.87:443 CLOSE_WAIT 5751/java tcp6 0 0 10.30.170.65:52860 10.30.170.145:2550 ESTABLISHED 5751/java tcp6 0 0 10.30.170.65:44750 10.30.170.189:2550 ESTABLISHED 5751/java tcp6 0 0 10.30.170.65:2550 10.30.170.189:37412 ESTABLISHED 5751/java tcp6 0 0 10.30.170.65:52870 10.30.170.145:2550 ESTABLISHED 5751/java tcp6 0 0 10.30.170.65:2550 10.30.170.189:37422 ESTABLISHED 5751/java tcp6 0 0 10.30.170.65:2550 10.30.170.145:37652 ESTABLISHED 5751/java tcp6 0 0 127.0.0.1:57626 127.0.0.1:44444 TIME_WAIT - tcp6 0 0 127.0.0.1:40874 127.0.0.1:1099 TIME_WAIT - tcp6 0 0 10.30.170.65:2550 10.30.170.145:37646 ESTABLISHED 5751/java tcp6 0 0 10.30.170.65:44754 10.30.170.189:2550 ESTABLISHED 5751/java udp 0 0 10.30.170.65:68 10.30.170.2:67 ESTABLISHED - udp 0 0 0.0.0.0:111 0.0.0.0:* - udp 0 0 127.0.0.1:323 0.0.0.0:* - udp6 0 0 :::111 :::* - udp6 0 0 ::1:323 :::* - looking for "BindException: Address already in use" in log file looking for "server is unhealthy" in log file + '[' 2 == 0 ']' + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_3_IP + echo 'Execute the post startup script on controller 10.30.170.189' Execute the post startup script on controller 10.30.170.189 + scp /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/post-startup-script.sh 10.30.170.189:/tmp/ Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + ssh 10.30.170.189 'bash /tmp/post-startup-script.sh 3' Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. tcp6 0 0 :::8101 :::* LISTEN Waiting up to 3 minutes for controller to come up, checking every 5 seconds... 2024-04-22T04:15:15,758 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 205 - org.opendaylight.infrautils.ready-api - 6.0.6 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.] Controller is UP 2024-04-22T04:15:15,758 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 205 - org.opendaylight.infrautils.ready-api - 6.0.6 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.] Listing all open ports on controller system... Active Internet connections (servers and established) Proto Recv-Q Send-Q Local Address Foreign Address State PID/Program name tcp 0 0 0.0.0.0:111 0.0.0.0:* LISTEN - tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN - tcp 0 0 10.30.170.189:22 10.30.171.144:44882 ESTABLISHED - tcp 0 0 10.30.170.189:54372 199.204.45.87:443 TIME_WAIT - tcp6 0 0 :::111 :::* LISTEN - tcp6 0 0 :::5555 :::* LISTEN - tcp6 0 0 :::8181 :::* LISTEN 5755/java tcp6 0 0 10.30.170.189:2550 :::* LISTEN 5755/java tcp6 0 0 :::22 :::* LISTEN - tcp6 0 0 127.0.0.1:44444 :::* LISTEN 5755/java tcp6 0 0 :::6653 :::* LISTEN 5755/java tcp6 0 0 :::8101 :::* LISTEN 5755/java tcp6 0 0 :::6633 :::* LISTEN 5755/java tcp6 0 0 127.0.0.1:1099 :::* LISTEN 5755/java tcp6 0 0 127.0.0.1:38637 :::* LISTEN 5755/java tcp6 1656 0 10.30.170.189:2550 10.30.170.65:44754 ESTABLISHED 5755/java tcp6 0 0 127.0.0.1:45284 127.0.0.1:44444 TIME_WAIT - tcp6 0 0 10.30.170.189:37422 10.30.170.65:2550 ESTABLISHED 5755/java tcp6 0 0 10.30.170.189:2550 10.30.170.145:35788 ESTABLISHED 5755/java tcp6 0 0 10.30.170.189:37412 10.30.170.65:2550 ESTABLISHED 5755/java tcp6 0 0 10.30.170.189:2550 10.30.170.65:44750 ESTABLISHED 5755/java tcp6 0 0 10.30.170.189:42592 10.30.170.145:2550 ESTABLISHED 5755/java tcp6 0 0 10.30.170.189:2550 10.30.170.145:35782 ESTABLISHED 5755/java tcp6 0 0 10.30.170.189:42588 10.30.170.145:2550 ESTABLISHED 5755/java tcp6 0 0 127.0.0.1:34866 127.0.0.1:1099 TIME_WAIT - tcp6 32 0 10.30.170.189:49576 199.204.45.87:443 CLOSE_WAIT 5755/java tcp6 0 0 127.0.0.1:45268 127.0.0.1:44444 TIME_WAIT - udp 0 0 127.0.0.1:323 0.0.0.0:* - udp 0 0 10.30.170.189:68 10.30.170.2:67 ESTABLISHED - udp 0 0 0.0.0.0:111 0.0.0.0:* - udp6 0 0 ::1:323 :::* - udp6 0 0 :::111 :::* - (Not all processes could be identified, non-owned process info will not be shown, you would have to be root to see it all.) looking for "BindException: Address already in use" in log file looking for "server is unhealthy" in log file + '[' 0 == 0 ']' + seed_index=1 + dump_controller_threads ++ seq 1 3 + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_1_IP + echo 'Let'\''s take the karaf thread dump' Let's take the karaf thread dump + ssh 10.30.170.145 'sudo ps aux' Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. ++ grep org.apache.karaf.main.Main /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/ps_before.log ++ grep -v grep ++ tr -s ' ' ++ cut -f2 '-d ' + pid=5753 + echo 'karaf main: org.apache.karaf.main.Main, pid:5753' karaf main: org.apache.karaf.main.Main, pid:5753 + ssh 10.30.170.145 '/usr/lib/jvm/java-17-openjdk/bin/jstack -l 5753' Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_2_IP + echo 'Let'\''s take the karaf thread dump' Let's take the karaf thread dump + ssh 10.30.170.65 'sudo ps aux' Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. ++ grep org.apache.karaf.main.Main /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/ps_before.log ++ grep -v grep ++ tr -s ' ' ++ cut -f2 '-d ' + pid=5751 + echo 'karaf main: org.apache.karaf.main.Main, pid:5751' karaf main: org.apache.karaf.main.Main, pid:5751 + ssh 10.30.170.65 '/usr/lib/jvm/java-17-openjdk/bin/jstack -l 5751' Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_3_IP + echo 'Let'\''s take the karaf thread dump' Let's take the karaf thread dump + ssh 10.30.170.189 'sudo ps aux' Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. ++ grep org.apache.karaf.main.Main /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/ps_before.log ++ grep -v grep ++ cut -f2 '-d ' ++ tr -s ' ' + pid=5755 + echo 'karaf main: org.apache.karaf.main.Main, pid:5755' karaf main: org.apache.karaf.main.Main, pid:5755 + ssh 10.30.170.189 '/usr/lib/jvm/java-17-openjdk/bin/jstack -l 5755' Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + '[' 0 -gt 0 ']' + echo 'Generating controller variables...' Generating controller variables... ++ seq 1 3 + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_1_IP + odl_variables=' -v ODL_SYSTEM_1_IP:10.30.170.145' + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_2_IP + odl_variables=' -v ODL_SYSTEM_1_IP:10.30.170.145 -v ODL_SYSTEM_2_IP:10.30.170.65' + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_3_IP + odl_variables=' -v ODL_SYSTEM_1_IP:10.30.170.145 -v ODL_SYSTEM_2_IP:10.30.170.65 -v ODL_SYSTEM_3_IP:10.30.170.189' + echo 'Generating mininet variables...' Generating mininet variables... ++ seq 1 1 + for i in $(seq 1 "${NUM_TOOLS_SYSTEM}") + MININETIP=TOOLS_SYSTEM_1_IP + tools_variables=' -v TOOLS_SYSTEM_1_IP:10.30.170.53' + get_test_suites SUITES + local __suite_list=SUITES + echo 'Locating test plan to use...' Locating test plan to use... + testplan_filepath=/w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/testplans/openflowplugin-clustering-calcium.txt + '[' '!' -f /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/testplans/openflowplugin-clustering-calcium.txt ']' + testplan_filepath=/w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/testplans/openflowplugin-clustering.txt + '[' disabled '!=' disabled ']' + echo 'Changing the testplan path...' Changing the testplan path... + sed s:integration:/w/workspace/openflowplugin-csit-3node-clustering-only-calcium: /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/testplans/openflowplugin-clustering.txt + cat testplan.txt # Place the suites in run order: /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/010__Cluster_HA_Owner_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/020__Cluster_HA_Owner_Restart.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/030__Cluster_HA_Data_Recovery_Leader_Follower_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustered_Reconciliation/010_Group_Flows.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/010_Switch_Disconnect.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/020_Cluster_Node_Failure.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/030_Cluster_Sync_Problems.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Bug_Validation/9145.robot + '[' -z '' ']' ++ grep -E -v '(^[[:space:]]*#|^[[:space:]]*$)' testplan.txt ++ tr '\012' ' ' + suite_list='/w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/010__Cluster_HA_Owner_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/020__Cluster_HA_Owner_Restart.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/030__Cluster_HA_Data_Recovery_Leader_Follower_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustered_Reconciliation/010_Group_Flows.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/010_Switch_Disconnect.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/020_Cluster_Node_Failure.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/030_Cluster_Sync_Problems.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Bug_Validation/9145.robot ' + eval 'SUITES='\''/w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/010__Cluster_HA_Owner_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/020__Cluster_HA_Owner_Restart.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/030__Cluster_HA_Data_Recovery_Leader_Follower_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustered_Reconciliation/010_Group_Flows.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/010_Switch_Disconnect.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/020_Cluster_Node_Failure.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/030_Cluster_Sync_Problems.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Bug_Validation/9145.robot '\''' ++ SUITES='/w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/010__Cluster_HA_Owner_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/020__Cluster_HA_Owner_Restart.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/030__Cluster_HA_Data_Recovery_Leader_Follower_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustered_Reconciliation/010_Group_Flows.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/010_Switch_Disconnect.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/020_Cluster_Node_Failure.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/030_Cluster_Sync_Problems.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Bug_Validation/9145.robot ' + echo 'Starting Robot test suites /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/010__Cluster_HA_Owner_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/020__Cluster_HA_Owner_Restart.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/030__Cluster_HA_Data_Recovery_Leader_Follower_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustered_Reconciliation/010_Group_Flows.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/010_Switch_Disconnect.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/020_Cluster_Node_Failure.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/030_Cluster_Sync_Problems.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Bug_Validation/9145.robot ...' Starting Robot test suites /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/010__Cluster_HA_Owner_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/020__Cluster_HA_Owner_Restart.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/030__Cluster_HA_Data_Recovery_Leader_Follower_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustered_Reconciliation/010_Group_Flows.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/010_Switch_Disconnect.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/020_Cluster_Node_Failure.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/030_Cluster_Sync_Problems.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Bug_Validation/9145.robot ... + robot -N openflowplugin-clustering.txt --removekeywords wuks -e exclude -e skip_if_calcium -v BUNDLEFOLDER:karaf-0.21.0 -v BUNDLE_URL:https://nexus.opendaylight.org/content/repositories//autorelease-7746/org/opendaylight/integration/karaf/0.21.0/karaf-0.21.0.zip -v CONTROLLER:10.30.170.145 -v CONTROLLER1:10.30.170.65 -v CONTROLLER2:10.30.170.189 -v CONTROLLER_USER:jenkins -v JAVA_HOME:/usr/lib/jvm/java-17-openjdk -v JDKVERSION:openjdk17 -v JENKINS_WORKSPACE:/w/workspace/openflowplugin-csit-3node-clustering-only-calcium -v MININET:10.30.170.53 -v MININET1: -v MININET2: -v MININET_USER:jenkins -v NEXUSURL_PREFIX:https://nexus.opendaylight.org -v NUM_ODL_SYSTEM:3 -v NUM_TOOLS_SYSTEM:1 -v ODL_STREAM:calcium -v ODL_SYSTEM_IP:10.30.170.145 -v ODL_SYSTEM_1_IP:10.30.170.145 -v ODL_SYSTEM_2_IP:10.30.170.65 -v ODL_SYSTEM_3_IP:10.30.170.189 -v ODL_SYSTEM_USER:jenkins -v TOOLS_SYSTEM_IP:10.30.170.53 -v TOOLS_SYSTEM_1_IP:10.30.170.53 -v TOOLS_SYSTEM_USER:jenkins -v USER_HOME:/home/jenkins -v IS_KARAF_APPL:True -v WORKSPACE:/tmp -v ODL_OF_PLUGIN:lithium /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/010__Cluster_HA_Owner_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/020__Cluster_HA_Owner_Restart.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustering/030__Cluster_HA_Data_Recovery_Leader_Follower_Failover.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Clustered_Reconciliation/010_Group_Flows.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/010_Switch_Disconnect.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/020_Cluster_Node_Failure.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/EntityOwnership/030_Cluster_Sync_Problems.robot /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/suites/openflowplugin/Bug_Validation/9145.robot ============================================================================== openflowplugin-clustering.txt ============================================================================== openflowplugin-clustering.txt.Cluster HA Owner Failover :: Test suite for C... ============================================================================== Check Shards Status Before Fail :: Check Status for all shards in ... | PASS | ------------------------------------------------------------------------------ Start Mininet Multiple Connections :: Start mininet tree,2 with co... | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor Before Fail... | PASS | ------------------------------------------------------------------------------ Reconnect Extra Switches To Successors And Check OVS Connections :... | PASS | ------------------------------------------------------------------------------ Check Network Operational Information Before Fail :: Check devices... | PASS | ------------------------------------------------------------------------------ Add Configuration In Owner and Verify Before Fail :: Add Flow in O... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Owner and Verify Before Fail :: Modify Flo... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Owner and Verify Before Fail :: Delete Flo... | PASS | ------------------------------------------------------------------------------ Add Configuration In Successor and Verify Before Fail :: Add Flow ... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Successor and Verify Before Fail :: Modify... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Successor and Verify Before Fail :: Delete... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Owner and Verify Before Fail :: Add Flow in Owner ... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Owner and Verify Before Fail :: Delete Flow in ... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Successor and Verify Before Fail :: Add Flow in Su... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Successor and Verify Before Fail :: Delete Flow... | PASS | ------------------------------------------------------------------------------ Modify Network And Verify Before Fail :: Take a link down and veri... | PASS | ------------------------------------------------------------------------------ Restore Network And Verify Before Fail :: Take the link up and ver... | PASS | ------------------------------------------------------------------------------ Kill Owner Instance :: Kill Owner Instance and verify it is dead :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Shards Status After Fail :: Create original cluster list and... | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor After Fail ... | PASS | ------------------------------------------------------------------------------ Check Network Operational Information After Fail :: Check devices ... | PASS | ------------------------------------------------------------------------------ Add Configuration In Owner and Verify After Fail :: Add Flow in Ow... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Owner and Verify After Fail :: Modify Flow... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Owner and Verify After Fail :: Delete Flow... | PASS | ------------------------------------------------------------------------------ Add Configuration In Successor and Verify After Fail :: Add Flow i... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Successor and Verify After Fail :: Modify ... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Successor and Verify After Fail :: Delete ... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Owner and Verify After Fail :: Add Flow in Owner a... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Owner and Verify After Fail :: Delete Flow in O... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Successor and Verify After Fail :: Add Flow in Suc... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Successor and Verify After Fail :: Delete Flow ... | PASS | ------------------------------------------------------------------------------ Modify Network and Verify After Fail :: Take a link down and verif... | PASS | ------------------------------------------------------------------------------ Restore Network and Verify After Fail :: Take the link up and veri... | FAIL | Keyword 'ClusterManagement.Check_Item_Occurrence_Member_List_Or_All' failed after retrying for 10 seconds. The last error was: '{"network-topology:network-topology":{"topology":[{"topology-id":"flow:1","node":[{"node-id":"openflow:2","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']","termination-point":[{"tp-id":"openflow:2:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:LOCAL\']"},{"tp-id":"openflow:2:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:1\']"},{"tp-id":"openflow:2:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:2\']"},{"tp-id":"openflow:2:3","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:3\']"}]},{"node-id":"openflow:3","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']","termination-point":[{"tp-id":"openflow:3:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:LOCAL\']"},{"tp-id":"openflow:3:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:1\']"},{"tp-id":"openflow:3:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:2\']"},{"tp-id":"openflow:3:3","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:3\']"}]},{"node-id":"openflow:1","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']","termination-point":[{"tp-id":"openflow:1:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:LOCAL\']"},{"tp-id":"openflow:1:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:1\']"},{"tp-id":"openflow:1:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:2\']"}]}],"link":[{"link-id":"openflow:3:3","source":{"source-node":"openflow:3","source-tp":"openflow:3:3"},"destination":{"dest-tp":"openflow:1:2","dest-node":"openflow:1"}},{"link-id":"openflow:1:2","source":{"source-node":"openflow:1","source-tp":"openflow:1:2"},"destination":{"dest-tp":"openflow:3:3","dest-node":"openflow:3"}}]}]}}' contains 'openflow:1' 16 times, not 21 times. ------------------------------------------------------------------------------ Start Old Owner Instance :: Start old Owner Instance and verify it... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Shards Status After Recover :: Create original cluster list ... | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status After Recover :: Check Entity Owner Stat... | PASS | ------------------------------------------------------------------------------ Check Network Operational Information After Recover :: Check devic... | FAIL | Keyword 'ClusterManagement.Check_Item_Occurrence_Member_List_Or_All' failed after retrying for 5 seconds. The last error was: '{"network-topology:network-topology":{"topology":[{"topology-id":"flow:1","link":[{"link-id":"openflow:3:3","source":{"source-node":"openflow:3","source-tp":"openflow:3:3"},"destination":{"dest-tp":"openflow:1:2","dest-node":"openflow:1"}},{"link-id":"openflow:1:2","source":{"source-node":"openflow:1","source-tp":"openflow:1:2"},"destination":{"dest-tp":"openflow:3:3","dest-node":"openflow:3"}}],"node":[{"node-id":"openflow:2","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']","termination-point":[{"tp-id":"openflow:2:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:LOCAL\']"},{"tp-id":"openflow:2:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:1\']"},{"tp-id":"openflow:2:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:2\']"},{"tp-id":"openflow:2:3","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:3\']"}]},{"node-id":"openflow:3","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']","termination-point":[{"tp-id":"openflow:3:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:LOCAL\']"},{"tp-id":"openflow:3:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:1\']"},{"tp-id":"openflow:3:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:2\']"},{"tp-id":"openflow:3:3","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:3\']"}]},{"node-id":"openflow:1","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']","termination-point":[{"tp-id":"openflow:1:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:2\']"},{"tp-id":"openflow:1:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:LOCAL\']"},{"tp-id":"openflow:1:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:1\']"}]}]}]}}' contains 'openflow:1' 16 times, not 21 times. ------------------------------------------------------------------------------ Add Configuration In Owner and Verify After Recover :: Add Flow in... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Owner and Verify After Recover :: Modify F... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Owner and Verify After Recover :: Delete F... | PASS | ------------------------------------------------------------------------------ Add Configuration In Old Owner and Verify After Recover :: Add Flo... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Old Owner and Verify After Recover :: Modi... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Old Owner and Verify After Recover :: Dele... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Owner and Verify After Recover :: Add Flow in Owne... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Owner and Verify After Recover :: Delete Flow i... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Old Owner and Verify After Recover :: Add Flow in ... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Old Owner and Verify After Recover :: Delete Fl... | PASS | ------------------------------------------------------------------------------ Modify Network and Verify After Recover :: Take a link down and ve... | PASS | ------------------------------------------------------------------------------ Restore Network and Verify After Recover :: Take the link up and v... | FAIL | Keyword 'ClusterManagement.Check_Item_Occurrence_Member_List_Or_All' failed after retrying for 10 seconds. The last error was: '{"network-topology:network-topology":{"topology":[{"topology-id":"flow:1","link":[{"link-id":"openflow:3:3","source":{"source-node":"openflow:3","source-tp":"openflow:3:3"},"destination":{"dest-tp":"openflow:1:2","dest-node":"openflow:1"}},{"link-id":"openflow:1:2","source":{"source-node":"openflow:1","source-tp":"openflow:1:2"},"destination":{"dest-tp":"openflow:3:3","dest-node":"openflow:3"}}],"node":[{"node-id":"openflow:2","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']","termination-point":[{"tp-id":"openflow:2:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:LOCAL\']"},{"tp-id":"openflow:2:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:1\']"},{"tp-id":"openflow:2:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:2\']"},{"tp-id":"openflow:2:3","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:3\']"}]},{"node-id":"openflow:3","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']","termination-point":[{"tp-id":"openflow:3:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:LOCAL\']"},{"tp-id":"openflow:3:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:1\']"},{"tp-id":"openflow:3:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:2\']"},{"tp-id":"openflow:3:3","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:3\']"}]},{"node-id":"openflow:1","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']","termination-point":[{"tp-id":"openflow:1:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:2\']"},{"tp-id":"openflow:1:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:LOCAL\']"},{"tp-id":"openflow:1:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:1\']"}]}]}]}}' contains 'openflow:1' 16 times, not 21 times. ------------------------------------------------------------------------------ Stop Mininet and Exit :: Stop mininet and exit connection. | PASS | ------------------------------------------------------------------------------ Check No Network Operational Information :: Check device is not in... | PASS | ------------------------------------------------------------------------------ openflowplugin-clustering.txt.Cluster HA Owner Failover :: Test su... | FAIL | 51 tests, 48 passed, 3 failed ============================================================================== openflowplugin-clustering.txt.Cluster HA Owner Restart :: Test suite for Cl... ============================================================================== Check Shards Status Before Stop :: Check Status for all shards in ... | PASS | ------------------------------------------------------------------------------ Start Mininet Multiple Connections :: Start mininet tree,2 with co... | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor Before Stop... | PASS | ------------------------------------------------------------------------------ Reconnect Extra Switches To Successors And Check OVS Connections :... | PASS | ------------------------------------------------------------------------------ Check Network Operational Information Before Stop :: Check devices... | PASS | ------------------------------------------------------------------------------ Add Configuration In Owner and Verify Before Stop :: Add Flow in O... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Owner and Verify Before Stop :: Modify Flo... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Owner and Verify Before Stop :: Delete Flo... | PASS | ------------------------------------------------------------------------------ Add Configuration In Successor and Verify Before Stop :: Add Flow ... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Successor and Verify Before Stop :: Modify... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Successor and Verify Before Stop :: Delete... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Owner and Verify Before Stop :: Add Flow in Owner ... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Owner and Verify Before Stop :: Delete Flow in ... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Successor and Verify Before Stop :: Add Flow in Su... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Successor and Verify Before Stop :: Delete Flow... | PASS | ------------------------------------------------------------------------------ Modify Network And Verify Before Stop :: Take a link down and veri... | PASS | ------------------------------------------------------------------------------ Restore Network And Verify Before Stop :: Take the link up and ver... | PASS | ------------------------------------------------------------------------------ Stop Owner Instance :: Stop Owner Instance and verify it is dead :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Shards Status After Stop :: Create original cluster list and... | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor After Stop ... | PASS | ------------------------------------------------------------------------------ Check Network Operational Information After Stop :: Check devices ... | PASS | ------------------------------------------------------------------------------ Add Configuration In Owner and Verify After Stop :: Add Flow in Ow... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Owner and Verify After Stop :: Modify Flow... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Owner and Verify After Stop :: Delete Flow... | PASS | ------------------------------------------------------------------------------ Add Configuration In Successor and Verify After Stop :: Add Flow i... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Successor and Verify After Stop :: Modify ... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Successor and Verify After Stop :: Delete ... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Owner and Verify After Stop :: Add Flow in Owner a... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Owner and Verify After Stop :: Delete Flow in O... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Successor and Verify After Stop :: Add Flow in Suc... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Successor and Verify After Stop :: Delete Flow ... | PASS | ------------------------------------------------------------------------------ Modify Network and Verify After Stop :: Take a link down and verif... | PASS | ------------------------------------------------------------------------------ Restore Network and Verify After Stop :: Take the link up and veri... | PASS | ------------------------------------------------------------------------------ Start Old Owner Instance :: Start old Owner Instance and verify it... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Shards Status After Start :: Create original cluster list an... | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status After Start :: Check Entity Owner Status... | PASS | ------------------------------------------------------------------------------ Check Network Operational Information After Start :: Check devices... | PASS | ------------------------------------------------------------------------------ Add Configuration In Owner and Verify After Start :: Add Flow in O... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Owner and Verify After Start :: Modify Flo... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Owner and Verify After Start :: Delete Flo... | PASS | ------------------------------------------------------------------------------ Add Configuration In Old Owner and Verify After Start :: Add Flow ... | PASS | ------------------------------------------------------------------------------ Modify Configuration In Old Owner and Verify After Start :: Modify... | PASS | ------------------------------------------------------------------------------ Delete Configuration In Old Owner and Verify After Start :: Delete... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Owner and Verify After Start :: Add Flow in Owner ... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Owner and Verify After Start :: Delete Flow in ... | PASS | ------------------------------------------------------------------------------ Send RPC Add to Old Owner and Verify After Start :: Add Flow in Ow... | PASS | ------------------------------------------------------------------------------ Send RPC Delete to Old Owner and Verify After Start :: Delete Flow... | PASS | ------------------------------------------------------------------------------ Modify Network and Verify After Start :: Take a link down and veri... | FAIL | Keyword 'ClusterManagement.Check_Item_Occurrence_Member_List_Or_All' failed after retrying for 20 seconds. The last error was: '{"network-topology:network-topology":{"topology":[{"topology-id":"flow:1","link":[{"link-id":"openflow:2:3","source":{"source-node":"openflow:2","source-tp":"openflow:2:3"},"destination":{"dest-tp":"openflow:1:1","dest-node":"openflow:1"}},{"link-id":"openflow:3:3","source":{"source-node":"openflow:3","source-tp":"openflow:3:3"},"destination":{"dest-tp":"openflow:1:2","dest-node":"openflow:1"}},{"link-id":"openflow:1:1","source":{"source-node":"openflow:1","source-tp":"openflow:1:1"},"destination":{"dest-tp":"openflow:2:3","dest-node":"openflow:2"}},{"link-id":"openflow:1:2","source":{"source-node":"openflow:1","source-tp":"openflow:1:2"},"destination":{"dest-tp":"openflow:3:3","dest-node":"openflow:3"}}],"node":[{"node-id":"openflow:2","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']","termination-point":[{"tp-id":"openflow:2:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:LOCAL\']"},{"tp-id":"openflow:2:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:1\']"},{"tp-id":"openflow:2:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:2\']/node-connector[id=\'openflow:2:2\']"},{"tp... [ Message content over the limit has been removed. ] ...ndaylight-inventory:nodes/node[id=\'openflow:3\']","termination-point":[{"tp-id":"openflow:3:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:LOCAL\']"},{"tp-id":"openflow:3:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:1\']"},{"tp-id":"openflow:3:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:2\']"},{"tp-id":"openflow:3:3","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:3\']/node-connector[id=\'openflow:3:3\']"}]},{"node-id":"openflow:1","opendaylight-topology-inventory:inventory-node-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']","termination-point":[{"tp-id":"openflow:1:2","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:2\']"},{"tp-id":"openflow:1:LOCAL","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:LOCAL\']"},{"tp-id":"openflow:1:1","opendaylight-topology-inventory:inventory-node-connector-ref":"/opendaylight-inventory:nodes/node[id=\'openflow:1\']/node-connector[id=\'openflow:1:1\']"}]}]}]}}' contains 'openflow:1' 21 times, not 16 times. ------------------------------------------------------------------------------ Restore Network and Verify After Start :: Take the link up and ver... | PASS | ------------------------------------------------------------------------------ Stop Mininet and Exit :: Stop mininet and exit connection. | PASS | ------------------------------------------------------------------------------ Check No Network Operational Information :: Check device is not in... | FAIL | Keyword 'ClusterManagement.Check_Item_Occurrence_Member_List_Or_All' failed after retrying for 20 seconds. The last error was: '{"network-topology:network-topology":{"topology":[{"topology-id":"flow:1","link":[{"link-id":"openflow:2:3","source":{"source-node":"openflow:2","source-tp":"openflow:2:3"},"destination":{"dest-tp":"openflow:1:1","dest-node":"openflow:1"}},{"link-id":"openflow:1:1","source":{"source-node":"openflow:1","source-tp":"openflow:1:1"},"destination":{"dest-tp":"openflow:2:3","dest-node":"openflow:2"}}]}]}}' contains 'openflow' 10 times, not 0 times. ------------------------------------------------------------------------------ openflowplugin-clustering.txt.Cluster HA Owner Restart :: Test sui... | FAIL | 51 tests, 49 passed, 2 failed ============================================================================== openflowplugin-clustering.txt.Cluster HA Data Recovery Leader Follower Fail... ============================================================================== Check Shards Status Before Leader Restart :: Check Status for all ... | PASS | ------------------------------------------------------------------------------ Get inventory Leader Before Leader Restart :: Find leader in the i... | PASS | ------------------------------------------------------------------------------ Start Mininet Connect To Follower Node1 :: Start mininet with conn... | PASS | ------------------------------------------------------------------------------ Add Flows In Follower Node2 and Verify Before Leader Restart :: Ad... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Stop Mininet Connected To Follower Node1 and Exit :: Stop mininet ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Restart Leader From Cluster Node :: Stop Leader Node and Start it ... | PASS | ------------------------------------------------------------------------------ Get inventory Follower After Leader Restart :: Find new Followers ... | PASS | ------------------------------------------------------------------------------ Start Mininet Connect To Old Leader :: Start mininet with connecti... | PASS | ------------------------------------------------------------------------------ Verify Flows In Switch After Leader Restart :: Verify flows are in... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Stop Mininet Connected To Old Leader and Exit :: Stop mininet and ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Restart Follower Node2 :: Stop Follower Node2 and Start it Up, Ver... | PASS | ------------------------------------------------------------------------------ Get inventory Follower After Follower Restart :: Find Followers an... | PASS | ------------------------------------------------------------------------------ Start Mininet Connect To Leader :: Start mininet with connection t... | PASS | ------------------------------------------------------------------------------ Verify Flows In Switch After Follower Restart :: Verify flows are ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Stop Mininet Connected To Leader and Exit :: Stop mininet Connecte... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Restart Full Cluster :: Stop all Cluster Nodes and Start it Up All. | PASS | ------------------------------------------------------------------------------ Get inventory Status After Cluster Restart :: Find New Followers a... | PASS | ------------------------------------------------------------------------------ Start Mininet Connect To Follower Node2 After Cluster Restart :: S... | PASS | ------------------------------------------------------------------------------ Verify Flows In Switch After Cluster Restart :: Verify flows are i... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Delete Flows In Follower Node1 and Verify After Leader Restart :: ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Stop Mininet Connected To Follower Node2 and Exit After Cluster Re... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ openflowplugin-clustering.txt.Cluster HA Data Recovery Leader Foll... | PASS | 21 tests, 21 passed, 0 failed ============================================================================== /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/test/csit/libraries/VsctlListParser.py:61: SyntaxWarning: "is not" with a literal. Did you mean "!="? if ctl_ref is not "": openflowplugin-clustering.txt.010 Group Flows :: Switch connections and clu... ============================================================================== Add Groups And Flows :: Add 100 groups 1&2 and flows in every switch. | PASS | ------------------------------------------------------------------------------ Start Mininet Multiple Connections :: Start mininet linear with co... | PASS | ------------------------------------------------------------------------------ Check Linear Topology :: Check Linear Topology. | PASS | ------------------------------------------------------------------------------ Check Stats Are Not Frozen :: Check that duration flow stat is inc... | PASS | ------------------------------------------------------------------------------ Check Flows In Operational DS :: Check Flows in operational DS. | PASS | ------------------------------------------------------------------------------ Check Groups In Operational DS :: Check Groups in operational DS. | PASS | ------------------------------------------------------------------------------ Check Flows In Switch :: Check Flows in switch. | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor Before Fail... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Disconnect Mininet From Owner :: Disconnect mininet from the owner :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? 10.30.170.189 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor After Fail ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Switch Moves To New Master :: Check switch s1 is connected t... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Linear Topology After Disconnect :: Check Linear Topology. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Stats Are Not Frozen After Disconnect :: Check that duration... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Remove Flows And Groups After Mininet Is Disconnected :: Remove 1 ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Operational DS After Mininet Is Disconnected :: Che... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Groups In Operational DS After Mininet Is Disconnected :: Ch... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Switch After Mininet Is Disconnected :: Check Flows... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Reconnect Mininet To Owner :: Reconnect mininet to switch 1 owner. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? 10.30.170.189 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor After Recon... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Add Flows And Groups After Owner Reconnect :: Add 1 group type 1&2... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Stats Are Not Frozen After Owner Reconnect :: Check that dur... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows After Owner Reconnect In Operational DS :: Check Flows... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Groups After Owner Reconnect In Operational DS :: Check Grou... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows After Owner Reconnect In Switch :: Check Flows in switch. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Switches Generate Slave Connection :: Check switches are con... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Disconnect Mininet From Successor :: Disconnect mininet from the S... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? 10.30.170.65 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find New Owner and Successor After D... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Disconnect Mininet From Current Owner :: Disconnect mininet from t... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? 10.30.170.145 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Current Owner and Successor Aft... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Switch Moves To Current Master :: Check switch s1 is connect... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Linear Topology After Owner Disconnect :: Check Linear Topol... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Stats Are Not Frozen After Owner Disconnect :: Check that du... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Remove Flows And Groups After Owner Disconnected :: Remove 1 group... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Operational DS After Owner Disconnected :: Check Fl... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Groups In Operational DS After Owner Disconnected :: Check G... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Switch After Owner Disconnected :: Check Flows in s... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Disconnect Mininet From Cluster :: Disconnect Mininet from Cluster. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? 10.30.170.189 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check No Switches After Disconnect :: Check no switches in topology. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | FAIL | Keyword 'ClusterOpenFlow.Check No Switches On Member' failed after retrying for 30 seconds. The last error was: '{"network-topology:network-topology":{"topology":[{"topology-id":"flow:1","link":[{"link-id":"openflow:2:2","source":{"source-node":"openflow:2","source-tp":"openflow:2:2"},"destination":{"dest-tp":"openflow:1:2","dest-node":"openflow:1"}},{"link-id":"openflow:3:2","source":{"source-node":"openflow:3","source-tp":"openflow:3:2"},"destination":{"dest-tp":"openflow:2:3","dest-node":"openflow:2"}},{"link-id":"openflow:2:3","source":{"source-node":"openflow:2","source-tp":"openflow:2:3"},"destination":{"dest-tp":"openflow:3:2","dest-node":"openflow:3"}},{"link-id":"openflow:1:2","source":{"source-node":"openflow:1","source-tp":"openflow:1:2"},"destination":{"dest-tp":"openflow:2:2","dest-node":"openflow:2"}}]}]}}' contains 'openflow:1' ------------------------------------------------------------------------------ Check Switch Is Not Connected :: Check switch s1 is not connected ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Reconnect Mininet To Cluster :: Reconnect mininet to cluster by re... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? 10.30.170.145 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? 10.30.170.65 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? 10.30.170.189 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Linear Topology After Mininet Reconnects :: Check Linear Top... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Add Flows And Groups After Mininet Reconnects :: Add 1 group type ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Operational DS After Mininet Reconnects :: Check Fl... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Groups In Operational DS After Mininet Reconnects :: Check G... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Switch After Mininet Reconnects :: Check Flows in s... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor Before Owne... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Switch Generates Slave Connection Before Owner Stop :: Check... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Shards Status Before Owner Stop :: Check Status for all shar... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Stop Owner Instance :: Stop Owner Instance and verify it is shutdown :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Shards Status After Stop :: Check Status for all shards in O... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor After Stop ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Stats Are Not Frozen After Owner Stop :: Check that duration... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Remove Configuration In Owner and Verify After Owner Stop :: Remov... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows After Owner Stop In Operational DS :: Check Flows in O... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Groups After Owner Stop In Operational DS :: Check Groups in... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Switch After Owner Stop :: Check Flows in switch. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Start Old Owner Instance :: Start old Owner Instance and verify it... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor After Start... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Linear Topology After Owner Restart :: Check Linear Topology. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Add Configuration In Owner and Verify After Owner Restart :: Add 1... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Stats Are Not Frozen After Owner Restart :: Check that durat... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Operational DS After Owner Restart :: Check Flows i... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Groups In Operational DS After Owner Restart :: Check Groups... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Switch After Owner Restart :: Check Flows in switch. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Restart Cluster :: Stop and Start cluster. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Linear Topology After Controller Restarts :: Check Linear To... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Stats Are Not Frozen After Cluster Restart :: Check that dur... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Operational DS After Controller Restarts :: Check F... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Groups In Operational DS After Controller Restarts :: Check ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check Flows In Switch After Controller Restarts :: Check Flows in ... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Stop Mininet :: Stop Mininet. :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Check No Switches :: Check no switches in topology. | FAIL | Keyword 'ClusterOpenFlow.Check No Switches On Member' failed after retrying for 5 seconds. The last error was: '{"network-topology:network-topology":{"topology":[{"topology-id":"flow:1","link":[{"link-id":"openflow:2:2","source":{"source-node":"openflow:2","source-tp":"openflow:2:2"},"destination":{"dest-tp":"openflow:1:2","dest-node":"openflow:1"}},{"link-id":"openflow:2:3","source":{"source-node":"openflow:2","source-tp":"openflow:2:3"},"destination":{"dest-tp":"openflow:3:2","dest-node":"openflow:3"}},{"link-id":"openflow:3:2","source":{"source-node":"openflow:3","source-tp":"openflow:3:2"},"destination":{"dest-tp":"openflow:2:3","dest-node":"openflow:2"}},{"link-id":"openflow:1:2","source":{"source-node":"openflow:1","source-tp":"openflow:1:2"},"destination":{"dest-tp":"openflow:2:2","dest-node":"openflow:2"}}]}]}}' contains 'openflow:1' ------------------------------------------------------------------------------ openflowplugin-clustering.txt.010 Group Flows :: Switch connection... | FAIL | 72 tests, 70 passed, 2 failed ============================================================================== openflowplugin-clustering.txt.010 Switch Disconnect :: Test suite for entit... ============================================================================== Switches To Be Connected To All Nodes :: Initial check for correct... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Reconnecting Switch s1 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Switches Still Be Connected To All Nodes :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ openflowplugin-clustering.txt.010 Switch Disconnect :: Test suite ... | PASS | 3 tests, 3 passed, 0 failed ============================================================================== openflowplugin-clustering.txt.020 Cluster Node Failure :: Test suite for en... ============================================================================== Switches To Be Connected To All Nodes :: Initial check for correct... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Restarting Owner Of Switch s1 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Switches Still Be Connected To All Nodes :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ openflowplugin-clustering.txt.020 Cluster Node Failure :: Test sui... | PASS | 3 tests, 3 passed, 0 failed ============================================================================== openflowplugin-clustering.txt.030 Cluster Sync Problems :: Test suite for e... ============================================================================== Start Mininet To All Nodes | PASS | ------------------------------------------------------------------------------ Switches To Be Connected To All Nodes :: Initial check for correct... :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | PASS | ------------------------------------------------------------------------------ Isolating Owner Of Switch s1 :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | FAIL | This test fails due to https://bugs.opendaylight.org/show_bug.cgi?id=6177 Keyword 'Check_Cluster_Is_In_Sync' failed after retrying for 1 minute. The last error was: HTTPError: 401 Client Error: Unauthorized for url: http://10.30.170.145:8181/jolokia/read/org.opendaylight.controller:Category=ShardManager,name=shard-manager-config,type=DistributedConfigDatastore Also keyword teardown failed: Keyword 'Check_Cluster_Is_In_Sync' failed after retrying for 1 minute. The last error was: Several failures occurred: 1) HTTPError: 401 Client Error: Unauthorized for url: http://10.30.170.145:8181/jolokia/read/org.opendaylight.controller:Category=ShardManager,name=shard-manager-config,type=DistributedConfigDatastore 2) Resolving variable '${response.text}' failed: AttributeError: 'NoneType' object has no attribute 'text' 3) Resolving variable '${response.status_code}' failed: AttributeError: 'NoneType' object has no attribute 'status_code' 4) Resolving variable '${response.status_code}' failed: AttributeError: 'No... [ Message content over the limit has been removed. ] 10) Resolving variable '${response.text}' failed: AttributeError: 'NoneType' object has no attribute 'text' 11) Resolving variable '${response.status_code}' failed: AttributeError: 'NoneType' object has no attribute 'status_code' 12) Resolving variable '${response.status_code}' failed: AttributeError: 'NoneType' object has no attribute 'status_code' 13) Resolving variable '${response.text}' failed: AttributeError: 'NoneType' object has no attribute 'text' 14) Evaluating expression "json.loads('''None''')" failed: JSONDecodeError: Expecting value: line 1 column 1 (char 0) 15) TypeError: Expected argument 1 to be a dictionary, got None instead. 16) TypeError: Expected argument 1 to be a dictionary, got None instead. 17) Index 1 has incorrect status: None ------------------------------------------------------------------------------ Switches Still Be Connected To All Nodes :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | FAIL | This test fails due to https://bugs.opendaylight.org/show_bug.cgi?id=6177 Keyword 'Check All Switches Connected To All Cluster Nodes' failed after retrying 15 times. The last error was: 'False' should be true. ------------------------------------------------------------------------------ Stop Mininet And Verify No Owners :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? :1: SyntaxWarning: "is not" with a literal. Did you mean "!="? | FAIL | This test fails due to https://bugs.opendaylight.org/show_bug.cgi?id=6177 Keyword 'Check No Device Owners In Controller' failed after retrying 15 times. The last error was: "{'odl-entity-owners:output': {'entities': [{'type': 'org.opendaylight.mdsal.AsyncServiceCloseEntityType', 'name': 'yanglib-mdsal-writer', 'candidate-nodes': ['member-2'], 'owner-node': 'member-2'}, {'type': 'org.opendaylight.mdsal.AsyncServiceCloseEntityType', 'name': 'ofp-topology-manager', 'candidate-nodes': ['member-2'], 'owner-node': 'member-2'}, {'type': 'org.opendaylight.mdsal.ServiceEntityType', 'name': 'openflow:1', 'candidate-nodes': ['member-1']}, {'type': 'org.opendaylight.mdsal.ServiceEntityType', 'name': 'yanglib-mdsal-writer', 'candidate-nodes': ['member-1', 'member-2', 'member-3'], 'owner-node': 'member-2'}, {'type': 'org.opendaylight.mdsal.ServiceEntityType', 'name': 'ofp-topology-manager', 'candidate-nodes': ['member-1', 'member-2', 'member-3'], 'owner-node': 'member-2'}]}}" contains 'member' 13 times, not 6 times. ------------------------------------------------------------------------------ openflowplugin-clustering.txt.030 Cluster Sync Problems :: Test su... | FAIL | 5 tests, 2 passed, 3 failed ============================================================================== openflowplugin-clustering.txt.9145 :: Switch connections and cluster are re... ============================================================================== Start Mininet Multiple Connections :: Start mininet linear with co... | FAIL | Keyword 'OVSDB.Check OVS OpenFlow Connections' failed after retrying for 10 seconds. The last error was: 'a03ccead-6e8d-4f04-8b85-97efdb60396a\n Manager "ptcp:6644"\n Bridge "s1"\n Controller "tcp:10.30.170.145:6633"\n Controller "tcp:10.30.170.189:6633"\n is_connected: true\n Controller "tcp:10.30.170.65:6633"\n is_connected: true\n fail_mode: standalone\n Port "s1"\n Interface "s1"\n type: internal\n Port "s1-eth1"\n Interface "s1-eth1"\n Port "s1-eth2"\n Interface "s1-eth2"\n ovs_version: "2.8.11"' contains 'is_connected' 2 times, not 3 times. ------------------------------------------------------------------------------ Check Entity Owner Status And Find Owner and Successor :: Check En... | FAIL | This test fails due to https://bugs.opendaylight.org/show_bug.cgi?id=9145 Keyword 'ClusterManagement.Verify_Owner_And_Successors_For_Device' failed after retrying for 30 seconds. The last error was: Could not parse owner and candidates for device openflow:1 ------------------------------------------------------------------------------ Stop Mininet :: Stop Mininet. | PASS | ------------------------------------------------------------------------------ openflowplugin-clustering.txt.9145 :: Switch connections and clust... | FAIL | 3 tests, 1 passed, 2 failed ============================================================================== openflowplugin-clustering.txt | FAIL | 209 tests, 197 passed, 12 failed ============================================================================== Output: /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/output.xml Log: /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/log.html Report: /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/report.html + true + echo 'Examining the files in data/log and checking filesize' Examining the files in data/log and checking filesize + ssh 10.30.170.145 'ls -altr /tmp/karaf-0.21.0/data/log/' Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. total 62852 drwxrwxr-x. 2 jenkins jenkins 48 Apr 22 04:14 . drwxrwxr-x. 9 jenkins jenkins 149 Apr 22 04:15 .. -rw-rw-r--. 1 jenkins jenkins 3516 Apr 22 04:16 karaf_console.log -rw-rw-r--. 1 jenkins jenkins 31319986 Apr 22 04:43 karaf.log + ssh 10.30.170.145 'du -hs /tmp/karaf-0.21.0/data/log/*' Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. 4.0K /tmp/karaf-0.21.0/data/log/karaf_console.log 62M /tmp/karaf-0.21.0/data/log/karaf.log + ssh 10.30.170.65 'ls -altr /tmp/karaf-0.21.0/data/log/' Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. total 65540 drwxrwxr-x. 2 jenkins jenkins 48 Apr 22 04:14 . -rw-rw-r--. 1 jenkins jenkins 1720 Apr 22 04:14 karaf_console.log drwxrwxr-x. 9 jenkins jenkins 149 Apr 22 04:15 .. -rw-rw-r--. 1 jenkins jenkins 41018738 Apr 22 04:43 karaf.log + ssh 10.30.170.65 'du -hs /tmp/karaf-0.21.0/data/log/*' Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. 4.0K /tmp/karaf-0.21.0/data/log/karaf_console.log 64M /tmp/karaf-0.21.0/data/log/karaf.log + ssh 10.30.170.189 'ls -altr /tmp/karaf-0.21.0/data/log/' Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. total 65476 drwxrwxr-x. 2 jenkins jenkins 48 Apr 22 04:14 . -rw-rw-r--. 1 jenkins jenkins 1720 Apr 22 04:14 karaf_console.log drwxrwxr-x. 9 jenkins jenkins 149 Apr 22 04:15 .. -rw-rw-r--. 1 jenkins jenkins 40363334 Apr 22 04:43 karaf.log + ssh 10.30.170.189 'du -hs /tmp/karaf-0.21.0/data/log/*' Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. 4.0K /tmp/karaf-0.21.0/data/log/karaf_console.log 64M /tmp/karaf-0.21.0/data/log/karaf.log + set +e ++ seq 1 3 + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_1_IP + echo 'Let'\''s take the karaf thread dump again' Let's take the karaf thread dump again + ssh 10.30.170.145 'sudo ps aux' Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. ++ grep org.apache.karaf.main.Main /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/ps_after.log ++ tr -s ' ' ++ cut -f2 '-d ' ++ grep -v grep + pid=12945 + echo 'karaf main: org.apache.karaf.main.Main, pid:12945' karaf main: org.apache.karaf.main.Main, pid:12945 + ssh 10.30.170.145 '/usr/lib/jvm/java-17-openjdk/bin/jstack -l 12945' Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + echo 'killing karaf process...' killing karaf process... + ssh 10.30.170.145 bash -c 'ps axf | grep karaf | grep -v grep | awk '\''{print "kill -9 " $1}'\'' | sh' Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_2_IP + echo 'Let'\''s take the karaf thread dump again' Let's take the karaf thread dump again + ssh 10.30.170.65 'sudo ps aux' Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. ++ grep org.apache.karaf.main.Main /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/ps_after.log ++ tr -s ' ' ++ cut -f2 '-d ' ++ grep -v grep + pid=10611 + echo 'karaf main: org.apache.karaf.main.Main, pid:10611' karaf main: org.apache.karaf.main.Main, pid:10611 + ssh 10.30.170.65 '/usr/lib/jvm/java-17-openjdk/bin/jstack -l 10611' Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + echo 'killing karaf process...' killing karaf process... + ssh 10.30.170.65 bash -c 'ps axf | grep karaf | grep -v grep | awk '\''{print "kill -9 " $1}'\'' | sh' Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_3_IP + echo 'Let'\''s take the karaf thread dump again' Let's take the karaf thread dump again + ssh 10.30.170.189 'sudo ps aux' Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. ++ grep org.apache.karaf.main.Main /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/ps_after.log ++ tr -s ' ' ++ cut -f2 '-d ' ++ grep -v grep + pid=18027 + echo 'karaf main: org.apache.karaf.main.Main, pid:18027' karaf main: org.apache.karaf.main.Main, pid:18027 + ssh 10.30.170.189 '/usr/lib/jvm/java-17-openjdk/bin/jstack -l 18027' Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + echo 'killing karaf process...' killing karaf process... + ssh 10.30.170.189 bash -c 'ps axf | grep karaf | grep -v grep | awk '\''{print "kill -9 " $1}'\'' | sh' Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + sleep 5 ++ seq 1 3 + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_1_IP + echo 'Compressing karaf.log 1' Compressing karaf.log 1 + ssh 10.30.170.145 gzip --best /tmp/karaf-0.21.0/data/log/karaf.log Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + echo 'Fetching compressed karaf.log 1' Fetching compressed karaf.log 1 + scp 10.30.170.145:/tmp/karaf-0.21.0/data/log/karaf.log.gz odl1_karaf.log.gz Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + ssh 10.30.170.145 rm -f /tmp/karaf-0.21.0/data/log/karaf.log.gz Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + scp 10.30.170.145:/tmp/karaf-0.21.0/data/log/karaf_console.log odl1_karaf_console.log Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + ssh 10.30.170.145 rm -f /tmp/karaf-0.21.0/data/log/karaf_console.log Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. + echo 'Fetch GC logs' Fetch GC logs + mkdir -p gclogs-1 + scp '10.30.170.145:/tmp/karaf-0.21.0/data/log/*.log' gclogs-1/ Warning: Permanently added '10.30.170.145' (ECDSA) to the list of known hosts. scp: /tmp/karaf-0.21.0/data/log/*.log: No such file or directory + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_2_IP + echo 'Compressing karaf.log 2' Compressing karaf.log 2 + ssh 10.30.170.65 gzip --best /tmp/karaf-0.21.0/data/log/karaf.log Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. gzip: /tmp/karaf-0.21.0/data/log/karaf.log: file size changed while zipping + echo 'Fetching compressed karaf.log 2' Fetching compressed karaf.log 2 + scp 10.30.170.65:/tmp/karaf-0.21.0/data/log/karaf.log.gz odl2_karaf.log.gz Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + ssh 10.30.170.65 rm -f /tmp/karaf-0.21.0/data/log/karaf.log.gz Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + scp 10.30.170.65:/tmp/karaf-0.21.0/data/log/karaf_console.log odl2_karaf_console.log Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + ssh 10.30.170.65 rm -f /tmp/karaf-0.21.0/data/log/karaf_console.log Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. + echo 'Fetch GC logs' Fetch GC logs + mkdir -p gclogs-2 + scp '10.30.170.65:/tmp/karaf-0.21.0/data/log/*.log' gclogs-2/ Warning: Permanently added '10.30.170.65' (ECDSA) to the list of known hosts. scp: /tmp/karaf-0.21.0/data/log/*.log: No such file or directory + for i in $(seq 1 "${NUM_ODL_SYSTEM}") + CONTROLLERIP=ODL_SYSTEM_3_IP + echo 'Compressing karaf.log 3' Compressing karaf.log 3 + ssh 10.30.170.189 gzip --best /tmp/karaf-0.21.0/data/log/karaf.log Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. gzip: /tmp/karaf-0.21.0/data/log/karaf.log: file size changed while zipping + echo 'Fetching compressed karaf.log 3' Fetching compressed karaf.log 3 + scp 10.30.170.189:/tmp/karaf-0.21.0/data/log/karaf.log.gz odl3_karaf.log.gz Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + ssh 10.30.170.189 rm -f /tmp/karaf-0.21.0/data/log/karaf.log.gz Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + scp 10.30.170.189:/tmp/karaf-0.21.0/data/log/karaf_console.log odl3_karaf_console.log Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + ssh 10.30.170.189 rm -f /tmp/karaf-0.21.0/data/log/karaf_console.log Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. + echo 'Fetch GC logs' Fetch GC logs + mkdir -p gclogs-3 + scp '10.30.170.189:/tmp/karaf-0.21.0/data/log/*.log' gclogs-3/ Warning: Permanently added '10.30.170.189' (ECDSA) to the list of known hosts. scp: /tmp/karaf-0.21.0/data/log/*.log: No such file or directory + echo 'Examine copied files' Examine copied files + ls -lt total 87844 drwxrwxr-x. 2 jenkins jenkins 6 Apr 22 04:44 gclogs-3 -rw-rw-r--. 1 jenkins jenkins 1720 Apr 22 04:44 odl3_karaf_console.log -rw-rw-r--. 1 jenkins jenkins 803350 Apr 22 04:44 odl3_karaf.log.gz drwxrwxr-x. 2 jenkins jenkins 6 Apr 22 04:44 gclogs-2 -rw-rw-r--. 1 jenkins jenkins 1720 Apr 22 04:44 odl2_karaf_console.log -rw-rw-r--. 1 jenkins jenkins 774189 Apr 22 04:44 odl2_karaf.log.gz drwxrwxr-x. 2 jenkins jenkins 6 Apr 22 04:44 gclogs-1 -rw-rw-r--. 1 jenkins jenkins 3516 Apr 22 04:44 odl1_karaf_console.log -rw-rw-r--. 1 jenkins jenkins 685920 Apr 22 04:44 odl1_karaf.log.gz -rw-rw-r--. 1 jenkins jenkins 137110 Apr 22 04:43 karaf_3_18027_threads_after.log -rw-rw-r--. 1 jenkins jenkins 15887 Apr 22 04:43 ps_after.log -rw-rw-r--. 1 jenkins jenkins 136537 Apr 22 04:43 karaf_2_10611_threads_after.log -rw-rw-r--. 1 jenkins jenkins 98364 Apr 22 04:43 karaf_1_12945_threads_after.log -rw-rw-r--. 1 jenkins jenkins 276865 Apr 22 04:43 report.html -rw-rw-r--. 1 jenkins jenkins 7680514 Apr 22 04:43 log.html -rw-rw-r--. 1 jenkins jenkins 78853450 Apr 22 04:43 output.xml -rw-rw-r--. 1 jenkins jenkins 1172 Apr 22 04:15 testplan.txt -rw-rw-r--. 1 jenkins jenkins 124670 Apr 22 04:15 karaf_3_5755_threads_before.log -rw-rw-r--. 1 jenkins jenkins 16458 Apr 22 04:15 ps_before.log -rw-rw-r--. 1 jenkins jenkins 120948 Apr 22 04:15 karaf_2_5751_threads_before.log -rw-rw-r--. 1 jenkins jenkins 123963 Apr 22 04:15 karaf_1_5753_threads_before.log -rw-rw-r--. 1 jenkins jenkins 3043 Apr 22 04:14 post-startup-script.sh -rw-rw-r--. 1 jenkins jenkins 1156 Apr 22 04:14 set_akka_debug.sh -rw-rw-r--. 1 jenkins jenkins 132 Apr 22 04:14 configplan.txt -rw-rw-r--. 1 jenkins jenkins 225 Apr 22 04:14 startup-script.sh -rw-rw-r--. 1 jenkins jenkins 3271 Apr 22 04:14 configuration-script.sh -rw-rw-r--. 1 jenkins jenkins 266 Apr 22 04:14 detect_variables.env -rw-rw-r--. 1 jenkins jenkins 86 Apr 22 04:14 set_variables.env -rw-rw-r--. 1 jenkins jenkins 356 Apr 22 04:14 slave_addresses.txt -rw-rw-r--. 1 jenkins jenkins 570 Apr 22 04:13 requirements.txt -rw-rw-r--. 1 jenkins jenkins 26 Apr 22 04:13 env.properties -rw-rw-r--. 1 jenkins jenkins 336 Apr 22 04:11 stack-parameters.yaml drwxrwxr-x. 7 jenkins jenkins 4096 Apr 22 04:10 test drwxrwxr-x. 2 jenkins jenkins 6 Apr 22 04:10 test@tmp + true [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/sh /tmp/jenkins9318785077801851490.sh Cleaning up Robot installation... $ ssh-agent -k unset SSH_AUTH_SOCK; unset SSH_AGENT_PID; echo Agent pid 5285 killed; [ssh-agent] Stopped. Recording plot data Robot results publisher started... INFO: Checking test criticality is deprecated and will be dropped in a future release! -Parsing output xml: Done! -Copying log files to build dir: Done! -Assigning results to build: Done! -Checking thresholds: Done! Done publishing Robot results. Build step 'Publish Robot Framework test results' changed build result to UNSTABLE [PostBuildScript] - [INFO] Executing post build scripts. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins11737431856260891417.sh Archiving csit artifacts mv: cannot stat '*_1.png': No such file or directory mv: cannot stat '/tmp/odl1_*': No such file or directory mv: cannot stat '*_2.png': No such file or directory mv: cannot stat '/tmp/odl2_*': No such file or directory mv: cannot stat '*_3.png': No such file or directory mv: cannot stat '/tmp/odl3_*': No such file or directory % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1685k 0 1685k 0 0 4355k 0 --:--:-- --:--:-- --:--:-- 4344k 100 7464k 0 7464k 0 0 5377k 0 --:--:-- 0:00:01 --:--:-- 5373k 100 9334k 0 9334k 0 0 5526k 0 --:--:-- 0:00:01 --:--:-- 5523k Archive: robot-plugin.zip inflating: ./archives/robot-plugin/log.html inflating: ./archives/robot-plugin/output.xml inflating: ./archives/robot-plugin/report.html mv: cannot stat '*.log.gz': No such file or directory mv: cannot stat '*.csv': No such file or directory mv: cannot stat '*.png': No such file or directory [PostBuildScript] - [INFO] Executing post build scripts. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins6304328391393101409.sh [PostBuildScript] - [INFO] Executing post build scripts. [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content OS_CLOUD=vex OS_STACK_NAME=releng-openflowplugin-csit-3node-clustering-only-calcium-227 [EnvInject] - Variables injected successfully. provisioning config files... copy managed file [clouds-yaml] to file:/home/jenkins/.config/openstack/clouds.yaml [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins15614740548090006051.sh ---> openstack-stack-delete.sh Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools[openstack] kubernetes python-heatclient python-openstackclient lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH INFO: Retrieving stack cost for: releng-openflowplugin-csit-3node-clustering-only-calcium-227 DEBUG: Successfully retrieved stack cost: total: 0.38999999999999996 INFO: Deleting stack releng-openflowplugin-csit-3node-clustering-only-calcium-227 Successfully deleted stack releng-openflowplugin-csit-3node-clustering-only-calcium-227 [PostBuildScript] - [INFO] Executing post build scripts. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins13515229903249956152.sh ---> sysstat.sh [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins8576985646320341487.sh ---> package-listing.sh ++ facter osfamily ++ tr '[:upper:]' '[:lower:]' + OS_FAMILY=redhat + workspace=/w/workspace/openflowplugin-csit-3node-clustering-only-calcium + START_PACKAGES=/tmp/packages_start.txt + END_PACKAGES=/tmp/packages_end.txt + DIFF_PACKAGES=/tmp/packages_diff.txt + PACKAGES=/tmp/packages_start.txt + '[' /w/workspace/openflowplugin-csit-3node-clustering-only-calcium ']' + PACKAGES=/tmp/packages_end.txt + case "${OS_FAMILY}" in + rpm -qa + sort + '[' -f /tmp/packages_start.txt ']' + '[' -f /tmp/packages_end.txt ']' + diff /tmp/packages_start.txt /tmp/packages_end.txt + '[' /w/workspace/openflowplugin-csit-3node-clustering-only-calcium ']' + mkdir -p /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/archives/ + cp -f /tmp/packages_diff.txt /tmp/packages_end.txt /tmp/packages_start.txt /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/archives/ [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins10452480159828460855.sh ---> capture-instance-metadata.sh Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH INFO: Running in OpenStack, capturing instance metadata [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins5928493328772574305.sh provisioning config files... Could not find credentials [logs] for openflowplugin-csit-3node-clustering-only-calcium #227 copy managed file [jenkins-log-archives-settings] to file:/w/workspace/openflowplugin-csit-3node-clustering-only-calcium@tmp/config8093027662670492289tmp Regular expression run condition: Expression=[^.*logs-s3.*], Label=[odl-logs-s3-cloudfront-index] Run condition [Regular expression match] enabling perform for step [Provide Configuration files] provisioning config files... copy managed file [jenkins-s3-log-ship] to file:/home/jenkins/.aws/credentials [EnvInject] - Injecting environment variables from a build step. [EnvInject] - Injecting as environment variables the properties content SERVER_ID=logs [EnvInject] - Variables injected successfully. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins9013242416771324172.sh ---> create-netrc.sh WARN: Log server credential not found. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins2024144343605945540.sh ---> python-tools-install.sh Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins12718766941872853589.sh ---> sudo-logs.sh Archiving 'sudo' log.. [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash /tmp/jenkins6528550169788479159.sh ---> job-cost.sh Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: zipp==1.1.0 python-openstackclient urllib3~=1.26.15 lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH DEBUG: total: 0.38999999999999996 INFO: Retrieving Stack Cost... INFO: Retrieving Pricing Info for: v3-standard-2 INFO: Archiving Costs [openflowplugin-csit-3node-clustering-only-calcium] $ /bin/bash -l /tmp/jenkins13340253862445508532.sh ---> logs-deploy.sh Setup pyenv: system 3.8.13 3.9.13 3.10.13 * 3.11.7 (set by /w/workspace/openflowplugin-csit-3node-clustering-only-calcium/.python-version) lf-activate-venv(): INFO: Reuse venv:/tmp/venv-z35k from file:/tmp/.os_lf_venv lf-activate-venv(): INFO: Installing: lftools lf-activate-venv(): INFO: Adding /tmp/venv-z35k/bin to PATH WARNING: Nexus logging server not set INFO: S3 path logs/releng/vex-yul-odl-jenkins-1/openflowplugin-csit-3node-clustering-only-calcium/227/ INFO: archiving logs to S3 ---> uname -a: Linux prd-centos8-robot-2c-8g-10017.novalocal 4.18.0-547.el8.x86_64 #1 SMP Wed Mar 20 00:35:01 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux ---> lscpu: Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 2 On-line CPU(s) list: 0,1 Thread(s) per core: 1 Core(s) per socket: 1 Socket(s): 2 NUMA node(s): 1 Vendor ID: AuthenticAMD CPU family: 23 Model: 49 Model name: AMD EPYC-Rome Processor Stepping: 0 CPU MHz: 2799.998 BogoMIPS: 5599.99 Virtualization: AMD-V Hypervisor vendor: KVM Virtualization type: full L1d cache: 32K L1i cache: 32K L2 cache: 512K L3 cache: 16384K NUMA node0 CPU(s): 0,1 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm rep_good nopl cpuid extd_apicid tsc_known_freq pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr wbnoinvd arat npt nrip_save umip rdpid arch_capabilities ---> nproc: 2 ---> df -h: Filesystem Size Used Avail Use% Mounted on devtmpfs 3.8G 0 3.8G 0% /dev tmpfs 3.8G 0 3.8G 0% /dev/shm tmpfs 3.8G 17M 3.8G 1% /run tmpfs 3.8G 0 3.8G 0% /sys/fs/cgroup /dev/vda1 40G 9.6G 31G 24% / tmpfs 770M 0 770M 0% /run/user/1001 ---> free -m: total used free shared buff/cache available Mem: 7697 624 4590 19 2482 6776 Swap: 1023 0 1023 ---> ip addr: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eth0: mtu 1458 qdisc mq state UP group default qlen 1000 link/ether fa:16:3e:4f:71:d3 brd ff:ff:ff:ff:ff:ff altname enp0s3 altname ens3 inet 10.30.171.144/23 brd 10.30.171.255 scope global dynamic noprefixroute eth0 valid_lft 84208sec preferred_lft 84208sec inet6 fe80::f816:3eff:fe4f:71d3/64 scope link valid_lft forever preferred_lft forever ---> sar -b -r -n DEV: Linux 4.18.0-547.el8.x86_64 (centos-stream-8-robot-8eca76b2-cb40-4fba-ac9e-845c8a03a7a7.noval) 04/22/2024 _x86_64_ (2 CPU) 04:09:06 LINUX RESTART (2 CPU) 04:10:00 AM tps rtps wtps bread/s bwrtn/s 04:10:01 AM 95.69 40.52 55.17 634.48 85029.31 04:11:01 AM 94.87 48.57 46.30 5989.20 13657.78 04:12:01 AM 109.58 6.68 102.90 1269.18 9402.75 04:13:01 AM 43.74 0.70 43.04 50.52 4627.23 04:14:01 AM 33.76 0.98 32.77 132.89 2052.43 04:15:01 AM 40.35 0.03 40.32 3.47 6768.18 04:16:01 AM 4.20 0.02 4.18 0.13 186.04 04:17:01 AM 2.15 0.07 2.08 3.20 217.08 04:18:01 AM 0.27 0.00 0.27 0.00 112.85 04:19:01 AM 0.22 0.00 0.22 0.00 98.67 04:20:01 AM 0.42 0.00 0.42 0.00 139.92 04:21:01 AM 0.38 0.00 0.38 0.00 144.51 04:22:01 AM 0.28 0.00 0.28 0.00 18.76 04:23:01 AM 0.25 0.00 0.25 0.00 144.39 04:24:01 AM 0.23 0.00 0.23 0.00 95.97 04:25:01 AM 1.45 0.03 1.42 0.27 90.10 04:26:01 AM 0.23 0.00 0.23 0.00 13.36 04:27:01 AM 0.22 0.00 0.22 0.00 17.48 04:28:01 AM 0.13 0.00 0.13 0.00 7.52 04:29:01 AM 0.23 0.00 0.23 0.00 131.39 04:30:01 AM 0.63 0.00 0.63 0.00 247.08 04:31:01 AM 0.15 0.00 0.15 0.00 35.90 04:32:01 AM 0.43 0.00 0.43 0.00 216.43 04:33:01 AM 0.30 0.00 0.30 0.00 228.87 04:34:01 AM 0.17 0.00 0.17 0.00 85.20 04:35:01 AM 0.62 0.00 0.62 0.00 20.49 04:36:01 AM 0.35 0.00 0.35 0.00 15.98 04:37:01 AM 0.38 0.00 0.38 0.00 56.04 04:38:01 AM 0.40 0.00 0.40 0.00 74.33 04:39:01 AM 0.30 0.00 0.30 0.00 140.96 04:40:01 AM 0.38 0.00 0.38 0.00 134.87 04:41:01 AM 0.37 0.00 0.37 0.00 190.20 04:42:01 AM 0.27 0.00 0.27 0.00 21.90 04:43:01 AM 0.27 0.00 0.27 0.00 36.14 04:44:01 AM 7.45 5.93 1.52 570.48 1352.28 04:45:01 AM 37.17 11.41 25.76 211.26 8834.14 Average: 10.98 2.15 8.83 235.39 1463.84 04:10:00 AM kbmemfree kbavail kbmemused %memused kbbuffers kbcached kbcommit %commit kbactive kbinact kbdirty 04:10:01 AM 6151664 7366388 1730764 21.96 2688 1421316 366304 4.65 110868 1399620 301944 04:11:01 AM 5453424 7034064 2429004 30.82 2688 1772324 729196 8.16 191416 1951608 107300 04:12:01 AM 5229500 7038448 2652928 33.66 2688 1986464 679028 7.60 211564 2096988 83044 04:13:01 AM 5185388 7015204 2697040 34.22 2688 2007292 691780 7.75 231504 2120888 1092 04:14:01 AM 4996876 7027220 2885552 36.61 2688 2200284 679364 7.61 261160 2256484 170012 04:15:01 AM 5016576 7049956 2865852 36.36 2688 2203136 644692 7.22 263056 2228472 24 04:16:01 AM 4969328 7005072 2913100 36.96 2688 2205500 729268 8.17 263252 2275204 2348 04:17:01 AM 4963532 7002236 2918896 37.03 2688 2208460 745096 8.34 263452 2280620 592 04:18:01 AM 4960300 7001704 2922128 37.07 2688 2211160 745096 8.34 263720 2283560 4 04:19:01 AM 4955004 7000368 2927424 37.14 2688 2215192 745096 8.34 263784 2288820 1116 04:20:01 AM 4949996 6999356 2932432 37.20 2688 2219124 781320 8.75 263784 2293420 988 04:21:01 AM 4947684 7000720 2934744 37.23 2688 2222784 781320 8.75 264244 2296048 432 04:22:01 AM 4944960 7001384 2937468 37.27 2688 2226192 781320 8.75 264244 2298568 3344 04:23:01 AM 4942148 7001504 2940280 37.30 2688 2229104 760288 8.51 264244 2301736 2000 04:24:01 AM 4937632 6998384 2944796 37.36 2688 2230500 764464 8.56 264244 2305876 596 04:25:01 AM 4937280 6999224 2945148 37.36 2688 2231688 764464 8.56 264316 2306436 124 04:26:01 AM 4937316 6999628 2945112 37.36 2688 2232048 746792 8.36 264312 2306620 168 04:27:01 AM 4936980 6999644 2945448 37.37 2688 2232400 746792 8.36 264312 2306872 48 04:28:01 AM 4936580 6999784 2945848 37.37 2688 2232940 746792 8.36 264312 2307320 388 04:29:01 AM 4929116 6996000 2953312 37.47 2688 2236620 779144 8.72 264368 2314528 172 04:30:01 AM 4920112 6993996 2962316 37.58 2688 2243628 779940 8.73 264364 2323140 52 04:31:01 AM 4918224 6995364 2964204 37.61 2688 2246928 779940 8.73 264364 2324932 2324 04:32:01 AM 4914292 6997568 2968136 37.66 2688 2253008 779940 8.73 264364 2329396 2048 04:33:01 AM 4906888 6995804 2975540 37.75 2688 2258680 779940 8.73 264364 2336780 920 04:34:01 AM 4904124 6994700 2978304 37.78 2688 2260308 745268 8.34 264364 2339656 44 04:35:01 AM 4903808 6994748 2978620 37.79 2688 2260668 745268 8.34 264364 2340076 4 04:36:01 AM 4904648 6996124 2977780 37.78 2688 2261204 745268 8.34 264364 2339372 180 04:37:01 AM 4902416 6995912 2980012 37.81 2688 2263224 762476 8.54 264504 2341228 640 04:38:01 AM 4897108 6992928 2985320 37.87 2688 2265556 804464 9.01 264516 2346256 952 04:39:01 AM 4890668 6989716 2991760 37.95 2688 2268772 804464 9.01 264516 2352584 16 04:40:01 AM 4881172 6985512 3001256 38.08 2688 2274052 823424 9.22 264516 2361844 1392 04:41:01 AM 4875708 6984312 3006720 38.14 2688 2278320 823424 9.22 264516 2366920 40 04:42:01 AM 4874176 6983408 3008252 38.16 2688 2278956 823424 9.22 264516 2368880 116 04:43:01 AM 4871200 6981332 3011228 38.20 2688 2279844 823424 9.22 264516 2371916 40 04:44:01 AM 4780060 6988308 3102368 39.36 2688 2377680 716920 8.03 299576 2426072 41168 04:45:01 AM 4702416 6936316 3180012 40.34 2688 2409012 716616 8.02 629064 2199016 13440 Average: 4981342 7009509 2901086 36.80 2688 2200121 746162 8.35 266582 2269104 20531 04:10:00 AM IFACE rxpck/s txpck/s rxkB/s txkB/s rxcmp/s txcmp/s rxmcst/s %ifutil 04:10:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:10:01 AM eth0 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:11:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:11:01 AM eth0 283.67 173.43 1510.35 58.37 0.00 0.00 0.00 0.00 04:12:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:12:01 AM eth0 31.94 27.54 265.98 5.34 0.00 0.00 0.00 0.00 04:13:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:13:01 AM eth0 37.24 23.16 268.86 4.77 0.00 0.00 0.00 0.00 04:14:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:14:01 AM eth0 42.17 36.89 296.20 8.08 0.00 0.00 0.00 0.00 04:15:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:15:01 AM eth0 690.25 569.99 137.30 117.28 0.00 0.00 0.00 0.00 04:16:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:16:01 AM eth0 36.44 28.21 20.79 4.41 0.00 0.00 0.00 0.00 04:17:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:17:01 AM eth0 10.46 10.98 12.08 2.02 0.00 0.00 0.00 0.00 04:18:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:18:01 AM eth0 8.77 8.25 14.89 1.48 0.00 0.00 0.00 0.00 04:19:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:19:01 AM eth0 9.32 10.40 16.39 1.95 0.00 0.00 0.00 0.00 04:20:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:20:01 AM eth0 22.46 22.59 18.83 3.51 0.00 0.00 0.00 0.00 04:21:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:21:01 AM eth0 12.53 14.51 13.05 2.54 0.00 0.00 0.00 0.00 04:22:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:22:01 AM eth0 7.23 8.91 15.60 1.67 0.00 0.00 0.00 0.00 04:23:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:23:01 AM eth0 29.15 28.98 13.42 2.85 0.00 0.00 0.00 0.00 04:24:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:24:01 AM eth0 64.06 60.63 15.23 5.61 0.00 0.00 0.00 0.00 04:25:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:25:01 AM eth0 61.18 60.13 13.61 5.80 0.00 0.00 0.00 0.00 04:26:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:26:01 AM eth0 10.68 10.23 2.25 1.93 0.00 0.00 0.00 0.00 04:27:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:27:01 AM eth0 12.53 9.70 2.43 1.81 0.00 0.00 0.00 0.00 04:28:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:28:01 AM eth0 7.48 6.98 2.72 1.09 0.00 0.00 0.00 0.00 04:29:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:29:01 AM eth0 56.20 53.90 12.67 13.17 0.00 0.00 0.00 0.00 04:30:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:30:01 AM eth0 156.35 143.32 104.64 10.74 0.00 0.00 0.00 0.00 04:31:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:31:01 AM eth0 93.39 83.81 45.06 6.71 0.00 0.00 0.00 0.00 04:32:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:32:01 AM eth0 98.58 85.94 71.60 7.33 0.00 0.00 0.00 0.00 04:33:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:33:01 AM eth0 59.53 59.83 67.25 5.26 0.00 0.00 0.00 0.00 04:34:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:34:01 AM eth0 34.59 34.48 23.63 6.69 0.00 0.00 0.00 0.00 04:35:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:35:01 AM eth0 11.35 11.08 2.40 2.18 0.00 0.00 0.00 0.00 04:36:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:36:01 AM eth0 7.82 7.52 2.88 1.51 0.00 0.00 0.00 0.00 04:37:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:37:01 AM eth0 61.84 56.72 30.20 4.89 0.00 0.00 0.00 0.00 04:38:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:38:01 AM eth0 64.46 42.62 17.57 4.24 0.00 0.00 0.00 0.00 04:39:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:39:01 AM eth0 13.81 14.71 15.56 2.16 0.00 0.00 0.00 0.00 04:40:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:40:01 AM eth0 55.38 46.22 30.87 4.86 0.00 0.00 0.00 0.00 04:41:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:41:01 AM eth0 7.70 8.90 19.63 1.51 0.00 0.00 0.00 0.00 04:42:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:42:01 AM eth0 2.65 2.75 0.74 0.52 0.00 0.00 0.00 0.00 04:43:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:43:01 AM eth0 36.48 20.80 7.39 2.22 0.00 0.00 0.00 0.00 04:44:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:44:01 AM eth0 109.63 86.37 670.10 7.40 0.00 0.00 0.00 0.00 04:45:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 04:45:01 AM eth0 160.01 119.51 987.01 211.95 0.00 0.00 0.00 0.00 Average: lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 Average: eth0 68.75 56.83 135.62 14.96 0.00 0.00 0.00 0.00 ---> sar -P ALL: Linux 4.18.0-547.el8.x86_64 (centos-stream-8-robot-8eca76b2-cb40-4fba-ac9e-845c8a03a7a7.noval) 04/22/2024 _x86_64_ (2 CPU) 04:09:06 LINUX RESTART (2 CPU) 04:10:00 AM CPU %user %nice %system %iowait %steal %idle 04:10:01 AM all 0.43 0.00 4.33 58.87 0.00 36.36 04:10:01 AM 0 0.00 0.00 4.35 22.61 0.00 73.04 04:10:01 AM 1 0.86 0.00 4.31 94.83 0.00 0.00 04:11:01 AM all 28.19 0.03 5.41 20.09 0.12 46.18 04:11:01 AM 0 30.74 0.03 4.66 18.93 0.10 45.53 04:11:01 AM 1 25.63 0.02 6.16 21.24 0.13 46.82 04:12:01 AM all 34.18 0.00 5.46 7.09 0.11 53.17 04:12:01 AM 0 14.56 0.00 3.96 7.08 0.10 74.30 04:12:01 AM 1 53.96 0.00 6.97 7.09 0.12 31.86 04:13:01 AM all 17.59 0.00 1.85 3.61 0.11 76.83 04:13:01 AM 0 9.23 0.00 1.76 4.01 0.12 84.88 04:13:01 AM 1 25.99 0.00 1.95 3.21 0.10 68.75 04:14:01 AM all 27.19 0.00 4.07 2.79 0.17 65.78 04:14:01 AM 0 15.17 0.00 3.50 2.12 0.12 79.09 04:14:01 AM 1 39.37 0.00 4.66 3.46 0.22 52.30 04:15:01 AM all 13.30 0.00 2.71 3.46 0.13 80.41 04:15:01 AM 0 9.60 0.00 2.45 2.26 0.13 85.56 04:15:01 AM 1 17.01 0.00 2.98 4.66 0.12 75.24 04:16:01 AM all 6.61 0.00 1.33 0.34 0.07 91.65 04:16:01 AM 0 5.77 0.00 1.10 0.00 0.07 93.06 04:16:01 AM 1 7.44 0.00 1.56 0.69 0.07 90.24 04:17:01 AM all 6.78 0.00 0.23 0.03 0.06 92.90 04:17:01 AM 0 8.80 0.00 0.27 0.00 0.05 90.88 04:17:01 AM 1 4.76 0.00 0.20 0.07 0.07 94.91 04:18:01 AM all 6.39 0.00 0.30 0.00 0.07 93.24 04:18:01 AM 0 8.17 0.00 0.42 0.00 0.05 91.36 04:18:01 AM 1 4.64 0.00 0.18 0.00 0.08 95.09 04:19:01 AM all 7.79 0.00 0.35 0.00 0.08 91.77 04:19:01 AM 0 2.54 0.00 0.25 0.00 0.07 97.14 04:19:01 AM 1 13.04 0.00 0.45 0.00 0.10 86.41 04:20:01 AM all 10.48 0.00 1.25 0.00 0.08 88.20 04:20:01 AM 0 9.39 0.00 0.59 0.00 0.08 89.94 04:20:01 AM 1 11.56 0.00 1.91 0.00 0.07 86.46 04:20:01 AM CPU %user %nice %system %iowait %steal %idle 04:21:01 AM all 8.79 0.00 0.37 0.00 0.08 90.77 04:21:01 AM 0 4.11 0.00 0.29 0.00 0.07 95.54 04:21:01 AM 1 13.43 0.00 0.45 0.00 0.08 86.04 04:22:01 AM all 8.10 0.00 0.41 0.00 0.08 91.41 04:22:01 AM 0 11.71 0.00 0.52 0.00 0.08 87.69 04:22:01 AM 1 4.53 0.00 0.30 0.00 0.07 95.11 04:23:01 AM all 6.87 0.00 1.28 0.03 0.07 91.76 04:23:01 AM 0 6.15 0.00 1.68 0.05 0.05 92.08 04:23:01 AM 1 7.59 0.00 0.89 0.00 0.08 91.44 04:24:01 AM all 7.57 0.00 0.68 0.03 0.08 91.64 04:24:01 AM 0 7.66 0.00 0.68 0.07 0.10 91.49 04:24:01 AM 1 7.48 0.00 0.67 0.00 0.07 91.78 04:25:01 AM all 7.78 0.00 0.74 0.19 0.08 91.21 04:25:01 AM 0 8.45 0.00 0.87 0.05 0.08 90.55 04:25:01 AM 1 7.12 0.00 0.62 0.32 0.07 91.87 04:26:01 AM all 4.35 0.00 0.34 0.00 0.09 95.21 04:26:01 AM 0 1.39 0.00 0.20 0.00 0.08 98.33 04:26:01 AM 1 7.32 0.00 0.49 0.00 0.10 92.09 04:27:01 AM all 4.42 0.00 0.33 0.00 0.08 95.16 04:27:01 AM 0 2.69 0.00 0.25 0.00 0.07 96.99 04:27:01 AM 1 6.16 0.00 0.42 0.00 0.10 93.32 04:28:01 AM all 4.62 0.00 0.30 0.00 0.09 94.99 04:28:01 AM 0 4.73 0.00 0.29 0.00 0.08 94.90 04:28:01 AM 1 4.52 0.00 0.30 0.00 0.10 95.08 04:29:01 AM all 10.33 0.00 0.72 0.05 0.09 88.81 04:29:01 AM 0 6.36 0.00 0.64 0.10 0.08 92.81 04:29:01 AM 1 14.37 0.00 0.81 0.00 0.09 84.74 04:30:01 AM all 8.85 0.00 0.91 0.03 0.11 90.10 04:30:01 AM 0 4.79 0.00 0.79 0.05 0.10 94.27 04:30:01 AM 1 12.95 0.00 1.04 0.00 0.12 85.90 04:31:01 AM all 6.65 0.00 1.02 0.00 0.10 92.23 04:31:01 AM 0 6.97 0.00 1.32 0.00 0.10 91.61 04:31:01 AM 1 6.34 0.00 0.72 0.00 0.10 92.83 04:31:01 AM CPU %user %nice %system %iowait %steal %idle 04:32:01 AM all 11.43 0.00 1.70 0.00 0.08 86.78 04:32:01 AM 0 7.89 0.00 1.71 0.00 0.08 90.32 04:32:01 AM 1 14.98 0.00 1.69 0.00 0.08 83.24 04:33:01 AM all 7.75 0.00 0.62 0.05 0.08 91.49 04:33:01 AM 0 3.32 0.00 0.50 0.10 0.08 95.99 04:33:01 AM 1 12.21 0.00 0.74 0.00 0.08 86.96 04:34:01 AM all 5.23 0.00 0.34 0.00 0.08 94.35 04:34:01 AM 0 2.53 0.00 0.30 0.00 0.08 97.09 04:34:01 AM 1 7.93 0.00 0.37 0.00 0.08 91.61 04:35:01 AM all 4.19 0.00 0.21 0.00 0.07 95.54 04:35:01 AM 0 2.83 0.00 0.20 0.00 0.07 96.90 04:35:01 AM 1 5.54 0.00 0.22 0.00 0.07 94.17 04:36:01 AM all 4.20 0.00 0.34 0.00 0.08 95.39 04:36:01 AM 0 1.24 0.00 0.15 0.00 0.07 98.55 04:36:01 AM 1 7.19 0.00 0.52 0.00 0.08 92.20 04:37:01 AM all 7.32 0.00 0.76 0.00 0.09 91.83 04:37:01 AM 0 3.84 0.00 0.47 0.00 0.08 95.60 04:37:01 AM 1 10.91 0.00 1.06 0.00 0.09 87.95 04:38:01 AM all 8.35 0.00 1.96 0.00 0.08 89.61 04:38:01 AM 0 8.56 0.00 0.96 0.00 0.08 90.40 04:38:01 AM 1 8.14 0.00 2.99 0.00 0.07 88.80 04:39:01 AM all 8.78 0.00 1.61 0.02 0.06 89.53 04:39:01 AM 0 6.21 0.00 0.28 0.03 0.05 93.43 04:39:01 AM 1 11.41 0.00 2.96 0.00 0.07 85.57 04:40:01 AM all 12.54 0.00 0.86 0.00 0.06 86.54 04:40:01 AM 0 3.33 0.00 0.62 0.00 0.07 95.98 04:40:01 AM 1 21.78 0.00 1.11 0.00 0.05 77.07 04:41:01 AM all 8.72 0.00 0.43 0.03 0.07 90.75 04:41:01 AM 0 0.83 0.00 0.22 0.07 0.05 98.83 04:41:01 AM 1 16.61 0.00 0.64 0.00 0.08 82.67 04:42:01 AM all 1.98 0.00 0.34 0.00 0.06 97.62 04:42:01 AM 0 1.60 0.00 0.27 0.00 0.05 98.08 04:42:01 AM 1 2.36 0.00 0.42 0.00 0.07 97.16 04:42:01 AM CPU %user %nice %system %iowait %steal %idle 04:43:01 AM all 3.66 0.00 0.35 0.00 0.08 95.91 04:43:01 AM 0 4.14 0.00 0.32 0.00 0.07 95.48 04:43:01 AM 1 3.18 0.00 0.38 0.00 0.08 96.35 04:44:01 AM all 14.02 6.11 1.10 0.44 0.08 78.25 04:44:01 AM 0 2.41 0.55 0.72 0.80 0.08 95.43 04:44:01 AM 1 25.63 11.67 1.47 0.08 0.07 61.08 04:45:01 AM all 26.72 10.96 3.43 2.30 0.08 56.51 04:45:01 AM 0 26.89 12.40 3.00 2.66 0.08 54.97 04:45:01 AM 1 26.55 9.51 3.86 1.94 0.08 58.06 Average: all 10.33 0.49 1.26 1.19 0.08 86.64 Average: 0 7.27 0.37 1.04 1.11 0.08 90.13 Average: 1 13.40 0.61 1.49 1.27 0.09 83.14