Started by upstream project "integration-distribution-mri-test-scandium" build number 76
originally caused by:
Started by timer
Running as SYSTEM
[EnvInject] - Loading node environment variables.
Building remotely on prd-centos8-robot-2c-8g-1207 (centos8-robot-2c-8g) in workspace /w/workspace/controller-csit-3node-clustering-tell-all-scandium
[ssh-agent] Looking for ssh-agent implementation...
[ssh-agent] Exec ssh-agent (binary ssh-agent on a remote machine)
$ ssh-agent
SSH_AUTH_SOCK=/tmp/ssh-3xzquIcrTk1n/agent.5162
SSH_AGENT_PID=5163
[ssh-agent] Started.
Running ssh-add (command line suppressed)
Identity added: /w/workspace/controller-csit-3node-clustering-tell-all-scandium@tmp/private_key_5906074262850596052.key (/w/workspace/controller-csit-3node-clustering-tell-all-scandium@tmp/private_key_5906074262850596052.key)
[ssh-agent] Using credentials jenkins (Release Engineering Jenkins Key)
The recommended git tool is: NONE
using credential opendaylight-jenkins-ssh
Wiping out workspace first.
Cloning the remote Git repository
Cloning repository git://devvexx.opendaylight.org/mirror/integration/test
> git init /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test # timeout=10
Fetching upstream changes from git://devvexx.opendaylight.org/mirror/integration/test
> git --version # timeout=10
> git --version # 'git version 2.43.0'
using GIT_SSH to set credentials Release Engineering Jenkins Key
[INFO] Currently running in a labeled security context
[INFO] Currently SELinux is 'enforcing' on the host
> /usr/bin/chcon --type=ssh_home_t /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test@tmp/jenkins-gitclient-ssh12468917035044533388.key
Verifying host key using known hosts file
You're using 'Known hosts file' strategy to verify ssh host keys, but your known_hosts file does not exist, please go to 'Manage Jenkins' -> 'Security' -> 'Git Host Key Verification Configuration' and configure host key verification.
> git fetch --tags --force --progress -- git://devvexx.opendaylight.org/mirror/integration/test +refs/heads/*:refs/remotes/origin/* # timeout=10
> git config remote.origin.url git://devvexx.opendaylight.org/mirror/integration/test # timeout=10
> git config --add remote.origin.fetch +refs/heads/*:refs/remotes/origin/* # timeout=10
> git config remote.origin.url git://devvexx.opendaylight.org/mirror/integration/test # timeout=10
Fetching upstream changes from git://devvexx.opendaylight.org/mirror/integration/test
using GIT_SSH to set credentials Release Engineering Jenkins Key
[INFO] Currently running in a labeled security context
[INFO] Currently SELinux is 'enforcing' on the host
> /usr/bin/chcon --type=ssh_home_t /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test@tmp/jenkins-gitclient-ssh9656849037753205002.key
Verifying host key using known hosts file
You're using 'Known hosts file' strategy to verify ssh host keys, but your known_hosts file does not exist, please go to 'Manage Jenkins' -> 'Security' -> 'Git Host Key Verification Configuration' and configure host key verification.
> git fetch --tags --force --progress -- git://devvexx.opendaylight.org/mirror/integration/test master # timeout=10
> git rev-parse FETCH_HEAD^{commit} # timeout=10
Checking out Revision 6c60ddfc8acc87c45ab0767b2ba1d2c4e7d34388 (origin/master)
> git config core.sparsecheckout # timeout=10
> git checkout -f 6c60ddfc8acc87c45ab0767b2ba1d2c4e7d34388 # timeout=10
Commit message: "Adapt test for new pce-allocation field"
> git rev-parse FETCH_HEAD^{commit} # timeout=10
> git rev-list --no-walk 62cb016f4f4171033927cf2ae7f4ac5095373e88 # timeout=10
No emails were triggered.
provisioning config files...
copy managed file [npmrc] to file:/home/jenkins/.npmrc
copy managed file [pipconf] to file:/home/jenkins/.config/pip/pip.conf
copy managed file [clouds-yaml] to file:/home/jenkins/.config/openstack/clouds.yaml
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins2007194133285548173.sh
---> python-tools-install.sh
Setup pyenv:
system
* 3.8.13 (set by /opt/pyenv/version)
* 3.9.13 (set by /opt/pyenv/version)
* 3.10.13 (set by /opt/pyenv/version)
* 3.11.7 (set by /opt/pyenv/version)
lf-activate-venv(): INFO: Creating python3 venv at /tmp/venv-DyLm
lf-activate-venv(): INFO: Save venv in file: /tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: lftools
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
Generating Requirements File
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
httplib2 0.31.0 requires pyparsing<4,>=3.0.4, but you have pyparsing 2.4.7 which is incompatible.
Python 3.11.7
pip 25.2 from /tmp/venv-DyLm/lib/python3.11/site-packages/pip (python 3.11)
appdirs==1.4.4
argcomplete==3.6.2
aspy.yaml==1.3.0
attrs==25.4.0
autopage==0.5.2
beautifulsoup4==4.14.2
boto3==1.40.45
botocore==1.40.45
bs4==0.0.2
cachetools==6.2.0
certifi==2025.10.5
cffi==2.0.0
cfgv==3.4.0
chardet==5.2.0
charset-normalizer==3.4.3
click==8.3.0
cliff==4.11.0
cmd2==2.7.0
cryptography==3.3.2
debtcollector==3.0.0
decorator==5.2.1
defusedxml==0.7.1
Deprecated==1.2.18
distlib==0.4.0
dnspython==2.8.0
docker==7.1.0
dogpile.cache==1.4.1
durationpy==0.10
email-validator==2.3.0
filelock==3.19.1
future==1.0.0
gitdb==4.0.12
GitPython==3.1.45
google-auth==2.41.1
httplib2==0.31.0
identify==2.6.15
idna==3.10
importlib-resources==1.5.0
iso8601==2.1.0
Jinja2==3.1.6
jmespath==1.0.1
jsonpatch==1.33
jsonpointer==3.0.0
jsonschema==4.25.1
jsonschema-specifications==2025.9.1
keystoneauth1==5.12.0
kubernetes==34.1.0
lftools==0.37.14
lxml==6.0.2
markdown-it-py==4.0.0
MarkupSafe==3.0.3
mdurl==0.1.2
msgpack==1.1.1
multi_key_dict==2.0.3
munch==4.0.0
netaddr==1.3.0
niet==1.4.2
nodeenv==1.9.1
oauth2client==4.1.3
oauthlib==3.3.1
openstacksdk==4.7.1
os-service-types==1.8.0
osc-lib==4.2.0
oslo.config==10.0.0
oslo.context==6.1.0
oslo.i18n==6.6.0
oslo.log==7.2.1
oslo.serialization==5.8.0
oslo.utils==9.1.0
packaging==25.0
pbr==7.0.1
platformdirs==4.4.0
prettytable==3.16.0
psutil==7.1.0
pyasn1==0.6.1
pyasn1_modules==0.4.2
pycparser==2.23
pygerrit2==2.0.15
PyGithub==2.8.1
Pygments==2.19.2
PyJWT==2.10.1
PyNaCl==1.6.0
pyparsing==2.4.7
pyperclip==1.11.0
pyrsistent==0.20.0
python-cinderclient==9.8.0
python-dateutil==2.9.0.post0
python-heatclient==4.3.0
python-jenkins==1.8.3
python-keystoneclient==5.7.0
python-magnumclient==4.9.0
python-openstackclient==8.2.0
python-swiftclient==4.8.0
PyYAML==6.0.3
referencing==0.36.2
requests==2.32.5
requests-oauthlib==2.0.0
requestsexceptions==1.4.0
rfc3986==2.0.0
rich==14.1.0
rich-argparse==1.7.1
rpds-py==0.27.1
rsa==4.9.1
ruamel.yaml==0.18.15
ruamel.yaml.clib==0.2.14
s3transfer==0.14.0
simplejson==3.20.2
six==1.17.0
smmap==5.0.2
soupsieve==2.8
stevedore==5.5.0
tabulate==0.9.0
toml==0.10.2
tomlkit==0.13.3
tqdm==4.67.1
typing_extensions==4.15.0
tzdata==2025.2
urllib3==1.26.20
virtualenv==20.34.0
wcwidth==0.2.14
websocket-client==1.8.0
wrapt==1.17.3
xdg==6.0.0
xmltodict==1.0.2
yq==3.4.3
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content
OS_STACK_TEMPLATE=csit-2-instance-type.yaml
OS_CLOUD=vex
OS_STACK_NAME=releng-controller-csit-3node-clustering-tell-all-scandium-76
OS_STACK_TEMPLATE_DIR=openstack-hot
[EnvInject] - Variables injected successfully.
provisioning config files...
copy managed file [clouds-yaml] to file:/home/jenkins/.config/openstack/clouds.yaml
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins5422865262728091250.sh
---> Create parameters file for OpenStack HOT
OpenStack Heat parameters generated
-----------------------------------
parameters:
vm_0_count: '3'
vm_0_flavor: 'v3-standard-4'
vm_0_image: 'ZZCI - Ubuntu 22.04 - builder - x86_64 - 20250201-010426.857'
vm_1_count: '0'
vm_1_flavor: 'v3-standard-2'
vm_1_image: 'ZZCI - Ubuntu 22.04 - mininet-ovs-217 - x86_64 - 20250201-060151.911'
job_name: '43164-76'
silo: 'releng'
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash -l /tmp/jenkins17808298517823761786.sh
---> Create HEAT stack
+ source /home/jenkins/lf-env.sh
+ lf-activate-venv --python python3 'lftools[openstack]' kubernetes niet python-heatclient python-openstackclient python-magnumclient urllib3~=1.26.15 yq
++ mktemp -d /tmp/venv-XXXX
+ lf_venv=/tmp/venv-xGHA
+ local venv_file=/tmp/.os_lf_venv
+ local python=python3
+ local options
+ local set_path=true
+ local install_args=
++ getopt -o np:v: -l no-path,system-site-packages,python:,venv-file: -n lf-activate-venv -- --python python3 'lftools[openstack]' kubernetes niet python-heatclient python-openstackclient python-magnumclient urllib3~=1.26.15 yq
+ options=' --python '\''python3'\'' -- '\''lftools[openstack]'\'' '\''kubernetes'\'' '\''niet'\'' '\''python-heatclient'\'' '\''python-openstackclient'\'' '\''python-magnumclient'\'' '\''urllib3~=1.26.15'\'' '\''yq'\'''
+ eval set -- ' --python '\''python3'\'' -- '\''lftools[openstack]'\'' '\''kubernetes'\'' '\''niet'\'' '\''python-heatclient'\'' '\''python-openstackclient'\'' '\''python-magnumclient'\'' '\''urllib3~=1.26.15'\'' '\''yq'\'''
++ set -- --python python3 -- 'lftools[openstack]' kubernetes niet python-heatclient python-openstackclient python-magnumclient urllib3~=1.26.15 yq
+ true
+ case $1 in
+ python=python3
+ shift 2
+ true
+ case $1 in
+ shift
+ break
+ case $python in
+ local pkg_list=
+ [[ -d /opt/pyenv ]]
+ echo 'Setup pyenv:'
Setup pyenv:
+ export PYENV_ROOT=/opt/pyenv
+ PYENV_ROOT=/opt/pyenv
+ export PATH=/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin
+ PATH=/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin
+ pyenv versions
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
+ command -v pyenv
++ pyenv init - --no-rehash
+ eval 'PATH="$(bash --norc -ec '\''IFS=:; paths=($PATH);
for i in ${!paths[@]}; do
if [[ ${paths[i]} == "'\'''\''/opt/pyenv/shims'\'''\''" ]]; then unset '\''\'\'''\''paths[i]'\''\'\'''\'';
fi; done;
echo "${paths[*]}"'\'')"
export PATH="/opt/pyenv/shims:${PATH}"
export PYENV_SHELL=bash
source '\''/opt/pyenv/libexec/../completions/pyenv.bash'\''
pyenv() {
local command
command="${1:-}"
if [ "$#" -gt 0 ]; then
shift
fi
case "$command" in
rehash|shell)
eval "$(pyenv "sh-$command" "$@")"
;;
*)
command pyenv "$command" "$@"
;;
esac
}'
+++ bash --norc -ec 'IFS=:; paths=($PATH);
for i in ${!paths[@]}; do
if [[ ${paths[i]} == "/opt/pyenv/shims" ]]; then unset '\''paths[i]'\'';
fi; done;
echo "${paths[*]}"'
++ PATH=/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin
++ export PATH=/opt/pyenv/shims:/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin
++ PATH=/opt/pyenv/shims:/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin
++ export PYENV_SHELL=bash
++ PYENV_SHELL=bash
++ source /opt/pyenv/libexec/../completions/pyenv.bash
+++ complete -F _pyenv pyenv
++ lf-pyver python3
++ local py_version_xy=python3
++ local py_version_xyz=
++ awk '{ print $1 }'
++ pyenv versions
++ local command
++ command=versions
++ '[' 1 -gt 0 ']'
++ shift
++ case "$command" in
++ command pyenv versions
++ pyenv versions
++ sed 's/^[ *]* //'
++ grep -E '^[0-9.]*[0-9]$'
++ [[ ! -s /tmp/.pyenv_versions ]]
+++ grep '^3' /tmp/.pyenv_versions
+++ tail -n 1
+++ sort -V
++ py_version_xyz=3.11.7
++ [[ -z 3.11.7 ]]
++ echo 3.11.7
++ return 0
+ pyenv local 3.11.7
+ local command
+ command=local
+ '[' 2 -gt 0 ']'
+ shift
+ case "$command" in
+ command pyenv local 3.11.7
+ pyenv local 3.11.7
+ for arg in "$@"
+ case $arg in
+ pkg_list+='lftools[openstack] '
+ for arg in "$@"
+ case $arg in
+ pkg_list+='kubernetes '
+ for arg in "$@"
+ case $arg in
+ pkg_list+='niet '
+ for arg in "$@"
+ case $arg in
+ pkg_list+='python-heatclient '
+ for arg in "$@"
+ case $arg in
+ pkg_list+='python-openstackclient '
+ for arg in "$@"
+ case $arg in
+ pkg_list+='python-magnumclient '
+ for arg in "$@"
+ case $arg in
+ pkg_list+='urllib3~=1.26.15 '
+ for arg in "$@"
+ case $arg in
+ pkg_list+='yq '
+ [[ -f /tmp/.os_lf_venv ]]
++ cat /tmp/.os_lf_venv
+ lf_venv=/tmp/venv-DyLm
+ echo 'lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from' file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
+ echo 'lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)'
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
+ local 'pip_opts=--upgrade --quiet'
+ pip_opts='--upgrade --quiet --trusted-host pypi.org'
+ pip_opts='--upgrade --quiet --trusted-host pypi.org --trusted-host files.pythonhosted.org'
+ pip_opts='--upgrade --quiet --trusted-host pypi.org --trusted-host files.pythonhosted.org --trusted-host pypi.python.org'
+ [[ -n '' ]]
+ [[ -n '' ]]
+ echo 'lf-activate-venv(): INFO: Attempting to install with network-safe options...'
lf-activate-venv(): INFO: Attempting to install with network-safe options...
+ /tmp/venv-DyLm/bin/python3 -m pip install --upgrade --quiet --trusted-host pypi.org --trusted-host files.pythonhosted.org --trusted-host pypi.python.org pip 'setuptools<66' virtualenv
+ echo 'lf-activate-venv(): INFO: Base packages installed successfully'
lf-activate-venv(): INFO: Base packages installed successfully
+ [[ -z lftools[openstack] kubernetes niet python-heatclient python-openstackclient python-magnumclient urllib3~=1.26.15 yq ]]
+ echo 'lf-activate-venv(): INFO: Installing additional packages: lftools[openstack] kubernetes niet python-heatclient python-openstackclient python-magnumclient urllib3~=1.26.15 yq '
lf-activate-venv(): INFO: Installing additional packages: lftools[openstack] kubernetes niet python-heatclient python-openstackclient python-magnumclient urllib3~=1.26.15 yq
+ /tmp/venv-DyLm/bin/python3 -m pip install --upgrade --quiet --trusted-host pypi.org --trusted-host files.pythonhosted.org --trusted-host pypi.python.org --upgrade-strategy eager 'lftools[openstack]' kubernetes niet python-heatclient python-openstackclient python-magnumclient urllib3~=1.26.15 yq
+ type python3
+ true
+ echo 'lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH'
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
+ PATH=/tmp/venv-DyLm/bin:/opt/pyenv/shims:/opt/pyenv/bin:/home/jenkins/.local/bin:/home/jenkins/bin:/usr/local/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/opt/puppetlabs/bin
+ return 0
+ openstack --os-cloud vex limits show --absolute
+--------------------------+---------+
| Name | Value |
+--------------------------+---------+
| maxTotalInstances | -1 |
| maxTotalCores | 450 |
| maxTotalRAMSize | 1000000 |
| maxServerMeta | 128 |
| maxImageMeta | 128 |
| maxPersonality | 5 |
| maxPersonalitySize | 10240 |
| maxTotalKeypairs | 100 |
| maxServerGroups | 10 |
| maxServerGroupMembers | 10 |
| maxTotalFloatingIps | -1 |
| maxSecurityGroups | -1 |
| maxSecurityGroupRules | -1 |
| totalRAMUsed | 778240 |
| totalCoresUsed | 190 |
| totalInstancesUsed | 64 |
| totalFloatingIpsUsed | 0 |
| totalSecurityGroupsUsed | 0 |
| totalServerGroupsUsed | 0 |
| maxTotalVolumes | -1 |
| maxTotalSnapshots | 10 |
| maxTotalVolumeGigabytes | 4096 |
| maxTotalBackups | 10 |
| maxTotalBackupGigabytes | 1000 |
| totalVolumesUsed | 0 |
| totalGigabytesUsed | 0 |
| totalSnapshotsUsed | 0 |
| totalBackupsUsed | 0 |
| totalBackupGigabytesUsed | 0 |
+--------------------------+---------+
+ pushd /opt/ciman/openstack-hot
/opt/ciman/openstack-hot /w/workspace/controller-csit-3node-clustering-tell-all-scandium
+ lftools openstack --os-cloud vex stack create releng-controller-csit-3node-clustering-tell-all-scandium-76 csit-2-instance-type.yaml /w/workspace/controller-csit-3node-clustering-tell-all-scandium/stack-parameters.yaml
Creating stack releng-controller-csit-3node-clustering-tell-all-scandium-76
Waiting to initialize infrastructure...
Stack initialization successful.
------------------------------------
Stack Details
------------------------------------
{'added': None,
'capabilities': [],
'created_at': '2025-10-06T19:29:03Z',
'deleted': None,
'deleted_at': None,
'description': 'No description',
'environment': None,
'environment_files': None,
'files': None,
'files_container': None,
'id': '632b8e55-fc07-4d8b-b9a6-ca8d5991c436',
'is_rollback_disabled': True,
'links': [{'href': 'https://orchestration.public.mtl1.vexxhost.net/v1/12c36e260d8e4bb2913965203b1b491f/stacks/releng-controller-csit-3node-clustering-tell-all-scandium-76/632b8e55-fc07-4d8b-b9a6-ca8d5991c436',
'rel': 'self'}],
'location': Munch({'cloud': 'vex', 'region_name': 'ca-ymq-1', 'zone': None, 'project': Munch({'id': '12c36e260d8e4bb2913965203b1b491f', 'name': '61975f2c-7c17-4d69-82fa-c3ae420ad6fd', 'domain_id': None, 'domain_name': 'Default'})}),
'name': 'releng-controller-csit-3node-clustering-tell-all-scandium-76',
'notification_topics': [],
'outputs': [{'description': 'IP addresses of the 1st vm types',
'output_key': 'vm_0_ips',
'output_value': ['10.30.170.103',
'10.30.171.180',
'10.30.171.153']},
{'description': 'IP addresses of the 2nd vm types',
'output_key': 'vm_1_ips',
'output_value': []}],
'owner_id': ****,
'parameters': {'OS::project_id': '12c36e260d8e4bb2913965203b1b491f',
'OS::stack_id': '632b8e55-fc07-4d8b-b9a6-ca8d5991c436',
'OS::stack_name': 'releng-controller-csit-3node-clustering-tell-all-scandium-76',
'job_name': '43164-76',
'silo': 'releng',
'vm_0_count': '3',
'vm_0_flavor': 'v3-standard-4',
'vm_0_image': 'ZZCI - Ubuntu 22.04 - builder - x86_64 - '
'20250201-010426.857',
'vm_1_count': '0',
'vm_1_flavor': 'v3-standard-2',
'vm_1_image': 'ZZCI - Ubuntu 22.04 - mininet-ovs-217 - x86_64 '
'- 20250201-060151.911'},
'parent_id': None,
'replaced': None,
'status': 'CREATE_COMPLETE',
'status_reason': 'Stack CREATE completed successfully',
'tags': [],
'template': None,
'template_description': 'No description',
'template_url': None,
'timeout_mins': 15,
'unchanged': None,
'updated': None,
'updated_at': None,
'user_project_id': 'eaf487173a50469ab7b263150be3a8a4'}
------------------------------------
+ popd
/w/workspace/controller-csit-3node-clustering-tell-all-scandium
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash -l /tmp/jenkins12811460903208966685.sh
---> Copy SSH public keys to CSIT lab
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: lftools[openstack] kubernetes python-heatclient python-openstackclient urllib3~=1.26.15
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
SSH not responding on 10.30.170.103. Retrying in 10 seconds...
SSH not responding on 10.30.171.180. Retrying in 10 seconds...
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
releng-43164-76-0-builder-2
Successfully copied public keys to slave 10.30.171.153
Ping to 10.30.170.103 successful.
Ping to 10.30.171.180 successful.
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
releng-43164-76-0-builder-0
Successfully copied public keys to slave 10.30.170.103
Process 6494 ready.
releng-43164-76-0-builder-1
Successfully copied public keys to slave 10.30.171.180
Process 6495 ready.
Process 6496 ready.
SSH ready on all stack servers.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash -l /tmp/jenkins7017975955935340812.sh
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Creating python3 venv at /tmp/venv-MHfE
lf-activate-venv(): INFO: Save venv in file: /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.robot_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: setuptools wheel
lf-activate-venv(): INFO: Adding /tmp/venv-MHfE/bin to PATH
+ echo 'Installing Python Requirements'
Installing Python Requirements
+ cat
+ python -m pip install -r requirements.txt
Looking in indexes: https://nexus3.opendaylight.org/repository/PyPi/simple
Collecting docker-py (from -r requirements.txt (line 1))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/docker-py/1.10.6/docker_py-1.10.6-py2.py3-none-any.whl (50 kB)
Collecting ipaddr (from -r requirements.txt (line 2))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/ipaddr/2.2.0/ipaddr-2.2.0.tar.gz (26 kB)
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Collecting netaddr (from -r requirements.txt (line 3))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/netaddr/1.3.0/netaddr-1.3.0-py3-none-any.whl (2.3 MB)
Collecting netifaces (from -r requirements.txt (line 4))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/netifaces/0.11.0/netifaces-0.11.0.tar.gz (30 kB)
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Collecting pyhocon (from -r requirements.txt (line 5))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/pyhocon/0.3.61/pyhocon-0.3.61-py3-none-any.whl (25 kB)
Collecting requests (from -r requirements.txt (line 6))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/requests/2.32.5/requests-2.32.5-py3-none-any.whl (64 kB)
Collecting robotframework (from -r requirements.txt (line 7))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework/7.3.2/robotframework-7.3.2-py3-none-any.whl (795 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 795.1/795.1 kB 26.9 MB/s 0:00:00
Collecting robotframework-httplibrary (from -r requirements.txt (line 8))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-httplibrary/0.4.2/robotframework-httplibrary-0.4.2.tar.gz (9.1 kB)
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Collecting robotframework-requests==0.9.7 (from -r requirements.txt (line 9))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-requests/0.9.7/robotframework_requests-0.9.7-py3-none-any.whl (21 kB)
Collecting robotframework-selenium2library (from -r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-selenium2library/3.0.0/robotframework_selenium2library-3.0.0-py2.py3-none-any.whl (6.2 kB)
Collecting robotframework-sshlibrary==3.8.0 (from -r requirements.txt (line 11))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-sshlibrary/3.8.0/robotframework-sshlibrary-3.8.0.tar.gz (51 kB)
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Collecting scapy (from -r requirements.txt (line 12))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/scapy/2.6.1/scapy-2.6.1-py3-none-any.whl (2.4 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.4/2.4 MB 39.1 MB/s 0:00:00
Collecting jsonpath-rw (from -r requirements.txt (line 15))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/jsonpath-rw/1.4.0/jsonpath-rw-1.4.0.tar.gz (13 kB)
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Collecting elasticsearch (from -r requirements.txt (line 18))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch/9.1.1/elasticsearch-9.1.1-py3-none-any.whl (937 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 937.5/937.5 kB 31.1 MB/s 0:00:00
Collecting elasticsearch-dsl (from -r requirements.txt (line 19))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch-dsl/8.18.0/elasticsearch_dsl-8.18.0-py3-none-any.whl (10 kB)
Collecting pyangbind (from -r requirements.txt (line 22))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/pyangbind/0.8.6/pyangbind-0.8.6-py3-none-any.whl (52 kB)
Collecting isodate (from -r requirements.txt (line 25))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/isodate/0.7.2/isodate-0.7.2-py3-none-any.whl (22 kB)
Collecting jmespath (from -r requirements.txt (line 28))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/jmespath/1.0.1/jmespath-1.0.1-py3-none-any.whl (20 kB)
Collecting jsonpatch (from -r requirements.txt (line 31))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/jsonpatch/1.33/jsonpatch-1.33-py2.py3-none-any.whl (12 kB)
Collecting paramiko>=1.15.3 (from robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/paramiko/4.0.0/paramiko-4.0.0-py3-none-any.whl (223 kB)
Collecting scp>=0.13.0 (from robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/scp/0.15.0/scp-0.15.0-py2.py3-none-any.whl (8.8 kB)
Collecting docker-pycreds>=0.2.1 (from docker-py->-r requirements.txt (line 1))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/docker-pycreds/0.4.0/docker_pycreds-0.4.0-py2.py3-none-any.whl (9.0 kB)
Collecting six>=1.4.0 (from docker-py->-r requirements.txt (line 1))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/six/1.17.0/six-1.17.0-py2.py3-none-any.whl (11 kB)
Collecting websocket-client>=0.32.0 (from docker-py->-r requirements.txt (line 1))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/websocket-client/1.8.0/websocket_client-1.8.0-py3-none-any.whl (58 kB)
Collecting pyparsing<4,>=2 (from pyhocon->-r requirements.txt (line 5))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/pyparsing/3.2.5/pyparsing-3.2.5-py3-none-any.whl (113 kB)
Collecting charset_normalizer<4,>=2 (from requests->-r requirements.txt (line 6))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/charset-normalizer/3.4.3/charset_normalizer-3.4.3-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (150 kB)
Collecting idna<4,>=2.5 (from requests->-r requirements.txt (line 6))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/idna/3.10/idna-3.10-py3-none-any.whl (70 kB)
Collecting urllib3<3,>=1.21.1 (from requests->-r requirements.txt (line 6))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/urllib3/2.5.0/urllib3-2.5.0-py3-none-any.whl (129 kB)
Collecting certifi>=2017.4.17 (from requests->-r requirements.txt (line 6))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/certifi/2025.10.5/certifi-2025.10.5-py3-none-any.whl (163 kB)
Collecting webtest>=2.0 (from robotframework-httplibrary->-r requirements.txt (line 8))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/webtest/3.0.6/webtest-3.0.6-py3-none-any.whl (32 kB)
Collecting jsonpointer (from robotframework-httplibrary->-r requirements.txt (line 8))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/jsonpointer/3.0.0/jsonpointer-3.0.0-py2.py3-none-any.whl (7.6 kB)
Collecting robotframework-seleniumlibrary>=3.0.0 (from robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-seleniumlibrary/6.8.0/robotframework_seleniumlibrary-6.8.0-py3-none-any.whl (104 kB)
Collecting ply (from jsonpath-rw->-r requirements.txt (line 15))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/ply/3.11/ply-3.11-py2.py3-none-any.whl (49 kB)
Collecting decorator (from jsonpath-rw->-r requirements.txt (line 15))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/decorator/5.2.1/decorator-5.2.1-py3-none-any.whl (9.2 kB)
Collecting elastic-transport<10,>=9.1.0 (from elasticsearch->-r requirements.txt (line 18))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elastic-transport/9.1.0/elastic_transport-9.1.0-py3-none-any.whl (65 kB)
Collecting python-dateutil (from elasticsearch->-r requirements.txt (line 18))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/python-dateutil/2.9.0.post0/python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB)
Collecting typing-extensions (from elasticsearch->-r requirements.txt (line 18))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/typing-extensions/4.15.0/typing_extensions-4.15.0-py3-none-any.whl (44 kB)
Collecting elasticsearch (from -r requirements.txt (line 18))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch/8.19.1/elasticsearch-8.19.1-py3-none-any.whl (940 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 940.5/940.5 kB 30.6 MB/s 0:00:00
INFO: pip is looking at multiple versions of elasticsearch-dsl to determine which version is compatible with other requirements. This could take a while.
Collecting elasticsearch-dsl (from -r requirements.txt (line 19))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch-dsl/8.17.1/elasticsearch_dsl-8.17.1-py3-none-any.whl (158 kB)
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch-dsl/8.17.0/elasticsearch_dsl-8.17.0-py3-none-any.whl (158 kB)
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch-dsl/8.16.0/elasticsearch_dsl-8.16.0-py3-none-any.whl (158 kB)
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elasticsearch-dsl/8.15.4/elasticsearch_dsl-8.15.4-py3-none-any.whl (104 kB)
Collecting elastic-transport<9,>=8.15.1 (from elasticsearch->-r requirements.txt (line 18))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/elastic-transport/8.17.1/elastic_transport-8.17.1-py3-none-any.whl (64 kB)
Collecting pyang (from pyangbind->-r requirements.txt (line 22))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/pyang/2.7.1/pyang-2.7.1-py2.py3-none-any.whl (598 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 598.5/598.5 kB 19.7 MB/s 0:00:00
Collecting lxml (from pyangbind->-r requirements.txt (line 22))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/lxml/6.0.2/lxml-6.0.2-cp311-cp311-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl (5.2 MB)
Collecting regex (from pyangbind->-r requirements.txt (line 22))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/regex/2025.9.18/regex-2025.9.18-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl (798 kB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 799.0/799.0 kB 27.3 MB/s 0:00:00
Collecting enum34 (from pyangbind->-r requirements.txt (line 22))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/enum34/1.1.10/enum34-1.1.10-py3-none-any.whl (11 kB)
Collecting bcrypt>=3.2 (from paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/bcrypt/5.0.0/bcrypt-5.0.0-cp39-abi3-manylinux_2_28_x86_64.whl (278 kB)
Collecting cryptography>=3.3 (from paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/cryptography/46.0.2/cryptography-46.0.2-cp311-abi3-manylinux_2_28_x86_64.whl (4.5 MB)
Collecting invoke>=2.0 (from paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/invoke/2.2.0/invoke-2.2.0-py3-none-any.whl (160 kB)
Collecting pynacl>=1.5 (from paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/pynacl/1.6.0/pynacl-1.6.0-cp38-abi3-manylinux_2_26_x86_64.manylinux_2_28_x86_64.whl (1.4 MB)
Collecting cffi>=2.0.0 (from cryptography>=3.3->paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/cffi/2.0.0/cffi-2.0.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (215 kB)
Collecting pycparser (from cffi>=2.0.0->cryptography>=3.3->paramiko>=1.15.3->robotframework-sshlibrary==3.8.0->-r requirements.txt (line 11))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/pycparser/2.23/pycparser-2.23-py3-none-any.whl (118 kB)
Collecting selenium>=4.3.0 (from robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/selenium/4.36.0/selenium-4.36.0-py3-none-any.whl (9.6 MB)
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 9.6/9.6 MB 35.3 MB/s 0:00:00
Collecting robotframework-pythonlibcore>=4.4.1 (from robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/robotframework-pythonlibcore/4.4.1/robotframework_pythonlibcore-4.4.1-py2.py3-none-any.whl (12 kB)
Collecting click>=8.0 (from robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/click/8.3.0/click-8.3.0-py3-none-any.whl (107 kB)
Collecting trio<1.0,>=0.30.0 (from selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/trio/0.31.0/trio-0.31.0-py3-none-any.whl (512 kB)
Collecting trio-websocket<1.0,>=0.12.2 (from selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/trio-websocket/0.12.2/trio_websocket-0.12.2-py3-none-any.whl (21 kB)
Collecting attrs>=23.2.0 (from trio<1.0,>=0.30.0->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/attrs/25.4.0/attrs-25.4.0-py3-none-any.whl (67 kB)
Collecting sortedcontainers (from trio<1.0,>=0.30.0->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/sortedcontainers/2.4.0/sortedcontainers-2.4.0-py2.py3-none-any.whl (29 kB)
Collecting outcome (from trio<1.0,>=0.30.0->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/outcome/1.3.0.post0/outcome-1.3.0.post0-py2.py3-none-any.whl (10 kB)
Collecting sniffio>=1.3.0 (from trio<1.0,>=0.30.0->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/sniffio/1.3.1/sniffio-1.3.1-py3-none-any.whl (10 kB)
Collecting wsproto>=0.14 (from trio-websocket<1.0,>=0.12.2->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/wsproto/1.2.0/wsproto-1.2.0-py3-none-any.whl (24 kB)
Collecting pysocks!=1.5.7,<2.0,>=1.5.6 (from urllib3[socks]<3.0,>=2.5.0->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/pysocks/1.7.1/PySocks-1.7.1-py3-none-any.whl (16 kB)
Collecting WebOb>=1.2 (from webtest>=2.0->robotframework-httplibrary->-r requirements.txt (line 8))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/webob/1.8.9/WebOb-1.8.9-py2.py3-none-any.whl (115 kB)
Collecting waitress>=3.0.2 (from webtest>=2.0->robotframework-httplibrary->-r requirements.txt (line 8))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/waitress/3.0.2/waitress-3.0.2-py3-none-any.whl (56 kB)
Collecting beautifulsoup4 (from webtest>=2.0->robotframework-httplibrary->-r requirements.txt (line 8))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/beautifulsoup4/4.14.2/beautifulsoup4-4.14.2-py3-none-any.whl (106 kB)
Collecting h11<1,>=0.9.0 (from wsproto>=0.14->trio-websocket<1.0,>=0.12.2->selenium>=4.3.0->robotframework-seleniumlibrary>=3.0.0->robotframework-selenium2library->-r requirements.txt (line 10))
Downloading https://nexus3.opendaylight.org/repository/PyPi/packages/h11/0.16.0/h11-0.16.0-py3-none-any.whl (37 kB)
Collecting soupsieve>1.2 (from beautifulsoup4->webtest>=2.0->robotframework-httplibrary->-r requirements.txt (line 8))
Using cached https://nexus3.opendaylight.org/repository/PyPi/packages/soupsieve/2.8/soupsieve-2.8-py3-none-any.whl (36 kB)
Building wheels for collected packages: robotframework-sshlibrary, ipaddr, netifaces, robotframework-httplibrary, jsonpath-rw
DEPRECATION: Building 'robotframework-sshlibrary' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'robotframework-sshlibrary'. Discussion can be found at https://github.com/pypa/pip/issues/6334
Building wheel for robotframework-sshlibrary (setup.py): started
Building wheel for robotframework-sshlibrary (setup.py): finished with status 'done'
Created wheel for robotframework-sshlibrary: filename=robotframework_sshlibrary-3.8.0-py3-none-any.whl size=55205 sha256=1393c359cd6b6c23f5f4a58965b721d0571ab5ac54ca1e2a761c9ff13e5d046c
Stored in directory: /home/jenkins/.cache/pip/wheels/f7/c9/b3/a977b7bcc410d45ae27d240df3d00a12585509180e373ecccc
DEPRECATION: Building 'ipaddr' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'ipaddr'. Discussion can be found at https://github.com/pypa/pip/issues/6334
Building wheel for ipaddr (setup.py): started
Building wheel for ipaddr (setup.py): finished with status 'done'
Created wheel for ipaddr: filename=ipaddr-2.2.0-py3-none-any.whl size=18353 sha256=1c37cc340476054efca833ce9a00135383116aff26c7127bb26ac9b152a90b60
Stored in directory: /home/jenkins/.cache/pip/wheels/dc/6c/04/da2d847fa8d45c59af3e1d83e2acc29cb8adcbaf04c0898dbf
DEPRECATION: Building 'netifaces' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'netifaces'. Discussion can be found at https://github.com/pypa/pip/issues/6334
Building wheel for netifaces (setup.py): started
Building wheel for netifaces (setup.py): finished with status 'done'
Created wheel for netifaces: filename=netifaces-0.11.0-cp311-cp311-linux_x86_64.whl size=41064 sha256=7cdf6dda26edd11bdd2ca64c07eec3e3e703daf806564f0091cbe10a91f4662c
Stored in directory: /home/jenkins/.cache/pip/wheels/f8/18/88/e61d54b995bea304bdb1d040a92b72228a1bf72ca2a3eba7c9
DEPRECATION: Building 'robotframework-httplibrary' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'robotframework-httplibrary'. Discussion can be found at https://github.com/pypa/pip/issues/6334
Building wheel for robotframework-httplibrary (setup.py): started
Building wheel for robotframework-httplibrary (setup.py): finished with status 'done'
Created wheel for robotframework-httplibrary: filename=robotframework_httplibrary-0.4.2-py3-none-any.whl size=10014 sha256=377a8493ef4baeb72ae57c2ff0fa3d4de1e5d78fae91d8d757d91e80d5257dd0
Stored in directory: /home/jenkins/.cache/pip/wheels/aa/bc/0d/9a20dd51effef392aae2733cb4c7b66c6fa29fca33d88b57ed
DEPRECATION: Building 'jsonpath-rw' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the `--use-pep517` option, (possibly combined with `--no-build-isolation`), or adding a `pyproject.toml` file to the source tree of 'jsonpath-rw'. Discussion can be found at https://github.com/pypa/pip/issues/6334
Building wheel for jsonpath-rw (setup.py): started
Building wheel for jsonpath-rw (setup.py): finished with status 'done'
Created wheel for jsonpath-rw: filename=jsonpath_rw-1.4.0-py3-none-any.whl size=15176 sha256=57163705aef2f07e7b2501c1589b43f53388c556fa248bb0a8f7f459dd114cd3
Stored in directory: /home/jenkins/.cache/pip/wheels/f1/54/63/9a8da38cefae13755097b36cc852decc25d8ef69c37d58d4eb
Successfully built robotframework-sshlibrary ipaddr netifaces robotframework-httplibrary jsonpath-rw
Installing collected packages: sortedcontainers, ply, netifaces, ipaddr, enum34, websocket-client, WebOb, waitress, urllib3, typing-extensions, soupsieve, sniffio, six, scapy, robotframework-pythonlibcore, robotframework, regex, pysocks, pyparsing, pycparser, netaddr, lxml, jsonpointer, jmespath, isodate, invoke, idna, h11, decorator, click, charset_normalizer, certifi, bcrypt, attrs, wsproto, requests, python-dateutil, pyhocon, pyang, outcome, jsonpath-rw, jsonpatch, elastic-transport, docker-pycreds, cffi, beautifulsoup4, webtest, trio, robotframework-requests, pynacl, pyangbind, elasticsearch, docker-py, cryptography, trio-websocket, robotframework-httplibrary, paramiko, elasticsearch-dsl, selenium, scp, robotframework-sshlibrary, robotframework-seleniumlibrary, robotframework-selenium2library
Successfully installed WebOb-1.8.9 attrs-25.4.0 bcrypt-5.0.0 beautifulsoup4-4.14.2 certifi-2025.10.5 cffi-2.0.0 charset_normalizer-3.4.3 click-8.3.0 cryptography-46.0.2 decorator-5.2.1 docker-py-1.10.6 docker-pycreds-0.4.0 elastic-transport-8.17.1 elasticsearch-8.19.1 elasticsearch-dsl-8.15.4 enum34-1.1.10 h11-0.16.0 idna-3.10 invoke-2.2.0 ipaddr-2.2.0 isodate-0.7.2 jmespath-1.0.1 jsonpatch-1.33 jsonpath-rw-1.4.0 jsonpointer-3.0.0 lxml-6.0.2 netaddr-1.3.0 netifaces-0.11.0 outcome-1.3.0.post0 paramiko-4.0.0 ply-3.11 pyang-2.7.1 pyangbind-0.8.6 pycparser-2.23 pyhocon-0.3.61 pynacl-1.6.0 pyparsing-3.2.5 pysocks-1.7.1 python-dateutil-2.9.0.post0 regex-2025.9.18 requests-2.32.5 robotframework-7.3.2 robotframework-httplibrary-0.4.2 robotframework-pythonlibcore-4.4.1 robotframework-requests-0.9.7 robotframework-selenium2library-3.0.0 robotframework-seleniumlibrary-6.8.0 robotframework-sshlibrary-3.8.0 scapy-2.6.1 scp-0.15.0 selenium-4.36.0 six-1.17.0 sniffio-1.3.1 sortedcontainers-2.4.0 soupsieve-2.8 trio-0.31.0 trio-websocket-0.12.2 typing-extensions-4.15.0 urllib3-2.5.0 waitress-3.0.2 websocket-client-1.8.0 webtest-3.0.6 wsproto-1.2.0
+ pip freeze
attrs==25.4.0
bcrypt==5.0.0
beautifulsoup4==4.14.2
certifi==2025.10.5
cffi==2.0.0
charset-normalizer==3.4.3
click==8.3.0
cryptography==46.0.2
decorator==5.2.1
distlib==0.4.0
docker-py==1.10.6
docker-pycreds==0.4.0
elastic-transport==8.17.1
elasticsearch==8.19.1
elasticsearch-dsl==8.15.4
enum34==1.1.10
filelock==3.19.1
h11==0.16.0
idna==3.10
invoke==2.2.0
ipaddr==2.2.0
isodate==0.7.2
jmespath==1.0.1
jsonpatch==1.33
jsonpath-rw==1.4.0
jsonpointer==3.0.0
lxml==6.0.2
netaddr==1.3.0
netifaces==0.11.0
outcome==1.3.0.post0
paramiko==4.0.0
platformdirs==4.4.0
ply==3.11
pyang==2.7.1
pyangbind==0.8.6
pycparser==2.23
pyhocon==0.3.61
PyNaCl==1.6.0
pyparsing==3.2.5
PySocks==1.7.1
python-dateutil==2.9.0.post0
regex==2025.9.18
requests==2.32.5
robotframework==7.3.2
robotframework-httplibrary==0.4.2
robotframework-pythonlibcore==4.4.1
robotframework-requests==0.9.7
robotframework-selenium2library==3.0.0
robotframework-seleniumlibrary==6.8.0
robotframework-sshlibrary==3.8.0
scapy==2.6.1
scp==0.15.0
selenium==4.36.0
six==1.17.0
sniffio==1.3.1
sortedcontainers==2.4.0
soupsieve==2.8
trio==0.31.0
trio-websocket==0.12.2
typing_extensions==4.15.0
urllib3==2.5.0
virtualenv==20.34.0
waitress==3.0.2
WebOb==1.8.9
websocket-client==1.8.0
WebTest==3.0.6
wsproto==1.2.0
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'env.properties'
[EnvInject] - Variables injected successfully.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash -l /tmp/jenkins3376056661184785336.sh
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: python-heatclient python-openstackclient yq
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
lftools 0.37.14 requires urllib3<2.1.0, but you have urllib3 2.5.0 which is incompatible.
kubernetes 34.1.0 requires urllib3<2.4.0,>=1.24.2, but you have urllib3 2.5.0 which is incompatible.
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
+ ODL_SYSTEM=()
+ TOOLS_SYSTEM=()
+ OPENSTACK_SYSTEM=()
+ OPENSTACK_CONTROLLERS=()
+ mapfile -t ADDR
++ openstack stack show -f json -c outputs releng-controller-csit-3node-clustering-tell-all-scandium-76
++ jq -r '.outputs[] | select(.output_key | match("^vm_[0-9]+_ips$")) | .output_value | .[]'
+ for i in "${ADDR[@]}"
++ ssh 10.30.170.103 hostname -s
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ REMHOST=releng-43164-76-0-builder-0
+ case ${REMHOST} in
+ ODL_SYSTEM=("${ODL_SYSTEM[@]}" "${i}")
+ for i in "${ADDR[@]}"
++ ssh 10.30.171.180 hostname -s
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ REMHOST=releng-43164-76-0-builder-1
+ case ${REMHOST} in
+ ODL_SYSTEM=("${ODL_SYSTEM[@]}" "${i}")
+ for i in "${ADDR[@]}"
++ ssh 10.30.171.153 hostname -s
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ REMHOST=releng-43164-76-0-builder-2
+ case ${REMHOST} in
+ ODL_SYSTEM=("${ODL_SYSTEM[@]}" "${i}")
+ echo NUM_ODL_SYSTEM=3
+ echo NUM_TOOLS_SYSTEM=0
+ '[' '' == yes ']'
+ NUM_OPENSTACK_SYSTEM=0
+ echo NUM_OPENSTACK_SYSTEM=0
+ '[' 0 -eq 2 ']'
+ echo ODL_SYSTEM_IP=10.30.170.103
++ seq 0 2
+ for i in $(seq 0 $(( ${#ODL_SYSTEM[@]} - 1 )))
+ echo ODL_SYSTEM_1_IP=10.30.170.103
+ for i in $(seq 0 $(( ${#ODL_SYSTEM[@]} - 1 )))
+ echo ODL_SYSTEM_2_IP=10.30.171.180
+ for i in $(seq 0 $(( ${#ODL_SYSTEM[@]} - 1 )))
+ echo ODL_SYSTEM_3_IP=10.30.171.153
+ echo TOOLS_SYSTEM_IP=
++ seq 0 -1
+ openstack_index=0
+ NUM_OPENSTACK_CONTROL_NODES=1
+ echo NUM_OPENSTACK_CONTROL_NODES=1
++ seq 0 0
+ for i in $(seq 0 $((NUM_OPENSTACK_CONTROL_NODES - 1)))
+ echo OPENSTACK_CONTROL_NODE_1_IP=
+ NUM_OPENSTACK_COMPUTE_NODES=-1
+ echo NUM_OPENSTACK_COMPUTE_NODES=-1
+ '[' -1 -ge 2 ']'
++ seq 0 -2
+ NUM_OPENSTACK_HAPROXY_NODES=0
+ echo NUM_OPENSTACK_HAPROXY_NODES=0
++ seq 0 -1
+ echo 'Contents of slave_addresses.txt:'
Contents of slave_addresses.txt:
+ cat slave_addresses.txt
NUM_ODL_SYSTEM=3
NUM_TOOLS_SYSTEM=0
NUM_OPENSTACK_SYSTEM=0
ODL_SYSTEM_IP=10.30.170.103
ODL_SYSTEM_1_IP=10.30.170.103
ODL_SYSTEM_2_IP=10.30.171.180
ODL_SYSTEM_3_IP=10.30.171.153
TOOLS_SYSTEM_IP=
NUM_OPENSTACK_CONTROL_NODES=1
OPENSTACK_CONTROL_NODE_1_IP=
NUM_OPENSTACK_COMPUTE_NODES=-1
NUM_OPENSTACK_HAPROXY_NODES=0
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'slave_addresses.txt'
[EnvInject] - Variables injected successfully.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/sh /tmp/jenkins8139767318640987735.sh
Preparing for JRE Version 21
Karaf artifact is karaf
Karaf project is integration
Java home is /usr/lib/jvm/java-21-openjdk-amd64
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'set_variables.env'
[EnvInject] - Variables injected successfully.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins3199212416196864723.sh
2025-10-06 19:30:53 URL:https://raw.githubusercontent.com/opendaylight/integration-distribution/stable/scandium/pom.xml [2619/2619] -> "pom.xml" [1]
Bundle version is 0.21.3-SNAPSHOT
--2025-10-06 19:30:53-- https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/maven-metadata.xml
Resolving nexus.opendaylight.org (nexus.opendaylight.org)... 199.204.45.87, 2604:e100:1:0:f816:3eff:fe45:48d6
Connecting to nexus.opendaylight.org (nexus.opendaylight.org)|199.204.45.87|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 1410 (1.4K) [application/xml]
Saving to: ‘maven-metadata.xml’
0K . 100% 47.0M=0s
2025-10-06 19:30:53 (47.0 MB/s) - ‘maven-metadata.xml’ saved [1410/1410]
org.opendaylight.integration
karaf
0.21.3-SNAPSHOT
20251006.183506
364
20251006183506
pom
0.21.3-20251006.183506-364
20251006183506
tar.gz
0.21.3-20251006.183506-364
20251006183506
zip
0.21.3-20251006.183506-364
20251006183506
cyclonedx
xml
0.21.3-20251006.183506-364
20251006183506
cyclonedx
json
0.21.3-20251006.183506-364
20251006183506
Nexus timestamp is 0.21.3-20251006.183506-364
Distribution bundle URL is https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
Distribution bundle is karaf-0.21.3-20251006.183506-364.zip
Distribution bundle version is 0.21.3-SNAPSHOT
Distribution folder is karaf-0.21.3-SNAPSHOT
Nexus prefix is https://nexus.opendaylight.org
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties file path 'detect_variables.env'
[EnvInject] - Variables injected successfully.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash -l /tmp/jenkins10340478701957982914.sh
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
lftools 0.37.14 requires urllib3<2.1.0, but you have urllib3 2.5.0 which is incompatible.
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: python-heatclient python-openstackclient
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
lftools 0.37.14 requires urllib3<2.1.0, but you have urllib3 2.5.0 which is incompatible.
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
Copying common-functions.sh to /tmp
Copying common-functions.sh to 10.30.170.103:/tmp
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
Copying common-functions.sh to 10.30.171.180:/tmp
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
Copying common-functions.sh to 10.30.171.153:/tmp
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins7190930483423804179.sh
common-functions.sh is being sourced
common-functions environment:
MAVENCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
ACTUALFEATURES:
FEATURESCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
CUSTOMPROP: /tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
LOGCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
MEMCONF: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
CONTROLLERMEM: 2048m
AKKACONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
MODULESCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
MODULESHARDSCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
SUITES:
#################################################
## Configure Cluster and Start ##
#################################################
ACTUALFEATURES: odl-integration-compatible-with-all,odl-jolokia,odl-restconf,odl-clustering-test-app
SPACE_SEPARATED_FEATURES: odl-integration-compatible-with-all odl-jolokia odl-restconf odl-clustering-test-app
Locating script plan to use...
script plan exists!!!
Changing the script plan path...
# Place the scripts in run order:
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/scripts/car_people_shard.sh
Executing /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/scripts/car_people_shard.sh...
Add car-people shards file
Copy shard config to member-1 with IP address 10.30.170.103
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
Copy shard config to member-2 with IP address 10.30.171.180
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
Copy shard config to member-3 with IP address 10.30.171.153
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
Finished running script plans
Configuring member-1 with IP address 10.30.170.103
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ source /tmp/common-functions.sh karaf-0.21.3-SNAPSHOT scandium
common-functions.sh is being sourced
++ [[ /tmp/common-functions.sh == \/\t\m\p\/\c\o\n\f\i\g\u\r\a\t\i\o\n\-\s\c\r\i\p\t\.\s\h ]]
++ echo 'common-functions.sh is being sourced'
++ BUNDLEFOLDER=karaf-0.21.3-SNAPSHOT
++ DISTROSTREAM=scandium
++ export MAVENCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
++ MAVENCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
++ export FEATURESCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
++ FEATURESCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
++ export CUSTOMPROP=/tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
++ CUSTOMPROP=/tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
++ export LOGCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
++ LOGCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
++ export MEMCONF=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
++ MEMCONF=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
++ export CONTROLLERMEM=
++ CONTROLLERMEM=
++ case "${DISTROSTREAM}" in
++ CLUSTER_SYSTEM=akka
++ export AKKACONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
++ AKKACONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
++ export MODULESCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
++ MODULESCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
++ export MODULESHARDSCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
++ MODULESHARDSCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
++ print_common_env
++ cat
common-functions environment:
MAVENCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
ACTUALFEATURES:
FEATURESCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
CUSTOMPROP: /tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
LOGCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
MEMCONF: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
CONTROLLERMEM:
AKKACONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
MODULESCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
MODULESHARDSCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
SUITES:
++ SSH='ssh -t -t'
++ extra_services_cntl=' dnsmasq.service httpd.service libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service rabbitmq-server.service '
++ extra_services_cmp=' libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service '
Changing to /tmp
+ echo 'Changing to /tmp'
+ cd /tmp
Downloading the distribution from https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
+ echo 'Downloading the distribution from https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip'
+ wget --progress=dot:mega https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
--2025-10-06 19:31:06-- https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
Resolving nexus.opendaylight.org (nexus.opendaylight.org)... 199.204.45.87, 2604:e100:1:0:f816:3eff:fe45:48d6
Connecting to nexus.opendaylight.org (nexus.opendaylight.org)|199.204.45.87|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 190794064 (182M) [application/zip]
Saving to: ‘karaf-0.21.3-20251006.183506-364.zip’
0K ........ ........ ........ ........ ........ ........ 1% 61.4M 3s
3072K ........ ........ ........ ........ ........ ........ 3% 107M 2s
6144K ........ ........ ........ ........ ........ ........ 4% 135M 2s
9216K ........ ........ ........ ........ ........ ........ 6% 140M 2s
12288K ........ ........ ........ ........ ........ ........ 8% 139M 2s
15360K ........ ........ ........ ........ ........ ........ 9% 126M 2s
18432K ........ ........ ........ ........ ........ ........ 11% 149M 1s
21504K ........ ........ ........ ........ ........ ........ 13% 160M 1s
24576K ........ ........ ........ ........ ........ ........ 14% 173M 1s
27648K ........ ........ ........ ........ ........ ........ 16% 177M 1s
30720K ........ ........ ........ ........ ........ ........ 18% 192M 1s
33792K ........ ........ ........ ........ ........ ........ 19% 233M 1s
36864K ........ ........ ........ ........ ........ ........ 21% 211M 1s
39936K ........ ........ ........ ........ ........ ........ 23% 218M 1s
43008K ........ ........ ........ ........ ........ ........ 24% 246M 1s
46080K ........ ........ ........ ........ ........ ........ 26% 249M 1s
49152K ........ ........ ........ ........ ........ ........ 28% 241M 1s
52224K ........ ........ ........ ........ ........ ........ 29% 252M 1s
55296K ........ ........ ........ ........ ........ ........ 31% 237M 1s
58368K ........ ........ ........ ........ ........ ........ 32% 231M 1s
61440K ........ ........ ........ ........ ........ ........ 34% 230M 1s
64512K ........ ........ ........ ........ ........ ........ 36% 228M 1s
67584K ........ ........ ........ ........ ........ ........ 37% 227M 1s
70656K ........ ........ ........ ........ ........ ........ 39% 228M 1s
73728K ........ ........ ........ ........ ........ ........ 41% 220M 1s
76800K ........ ........ ........ ........ ........ ........ 42% 230M 1s
79872K ........ ........ ........ ........ ........ ........ 44% 226M 1s
82944K ........ ........ ........ ........ ........ ........ 46% 230M 1s
86016K ........ ........ ........ ........ ........ ........ 47% 232M 1s
89088K ........ ........ ........ ........ ........ ........ 49% 225M 1s
92160K ........ ........ ........ ........ ........ ........ 51% 230M 0s
95232K ........ ........ ........ ........ ........ ........ 52% 231M 0s
98304K ........ ........ ........ ........ ........ ........ 54% 239M 0s
101376K ........ ........ ........ ........ ........ ........ 56% 227M 0s
104448K ........ ........ ........ ........ ........ ........ 57% 242M 0s
107520K ........ ........ ........ ........ ........ ........ 59% 228M 0s
110592K ........ ........ ........ ........ ........ ........ 61% 231M 0s
113664K ........ ........ ........ ........ ........ ........ 62% 243M 0s
116736K ........ ........ ........ ........ ........ ........ 64% 219M 0s
119808K ........ ........ ........ ........ ........ ........ 65% 239M 0s
122880K ........ ........ ........ ........ ........ ........ 67% 248M 0s
125952K ........ ........ ........ ........ ........ ........ 69% 248M 0s
129024K ........ ........ ........ ........ ........ ........ 70% 244M 0s
132096K ........ ........ ........ ........ ........ ........ 72% 253M 0s
135168K ........ ........ ........ ........ ........ ........ 74% 247M 0s
138240K ........ ........ ........ ........ ........ ........ 75% 237M 0s
141312K ........ ........ ........ ........ ........ ........ 77% 252M 0s
144384K ........ ........ ........ ........ ........ ........ 79% 237M 0s
147456K ........ ........ ........ ........ ........ ........ 80% 238M 0s
150528K ........ ........ ........ ........ ........ ........ 82% 242M 0s
153600K ........ ........ ........ ........ ........ ........ 84% 236M 0s
156672K ........ ........ ........ ........ ........ ........ 85% 243M 0s
159744K ........ ........ ........ ........ ........ ........ 87% 236M 0s
162816K ........ ........ ........ ........ ........ ........ 89% 239M 0s
165888K ........ ........ ........ ........ ........ ........ 90% 221M 0s
168960K ........ ........ ........ ........ ........ ........ 92% 203M 0s
172032K ........ ........ ........ ........ ........ ........ 93% 232M 0s
175104K ........ ........ ........ ........ ........ ........ 95% 242M 0s
178176K ........ ........ ........ ........ ........ ........ 97% 243M 0s
181248K ........ ........ ........ ........ ........ ........ 98% 227M 0s
184320K ........ ........ ........ ....... 100% 230M=0.9s
2025-10-06 19:31:07 (204 MB/s) - ‘karaf-0.21.3-20251006.183506-364.zip’ saved [190794064/190794064]
Extracting the new controller...
+ echo 'Extracting the new controller...'
+ unzip -q karaf-0.21.3-20251006.183506-364.zip
Adding external repositories...
+ echo 'Adding external repositories...'
+ sed -ie 's%org.ops4j.pax.url.mvn.repositories=%org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases%g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
#
# If set to true, the following property will not allow any certificate to be used
# when accessing Maven repositories through SSL
#
#org.ops4j.pax.url.mvn.certificateCheck=
#
# Path to the local Maven settings file.
# The repositories defined in this file will be automatically added to the list
# of default repositories if the 'org.ops4j.pax.url.mvn.repositories' property
# below is not set.
# The following locations are checked for the existence of the settings.xml file
# * 1. looks for the specified url
# * 2. if not found looks for ${user.home}/.m2/settings.xml
# * 3. if not found looks for ${maven.home}/conf/settings.xml
# * 4. if not found looks for ${M2_HOME}/conf/settings.xml
#
#org.ops4j.pax.url.mvn.settings=
#
# Path to the local Maven repository which is used to avoid downloading
# artifacts when they already exist locally.
# The value of this property will be extracted from the settings.xml file
# above, or defaulted to:
# System.getProperty( "user.home" ) + "/.m2/repository"
#
org.ops4j.pax.url.mvn.localRepository=${karaf.home}/${karaf.default.repository}
#
# Default this to false. It's just weird to use undocumented repos
#
org.ops4j.pax.url.mvn.useFallbackRepositories=false
#
# Uncomment if you don't wanna use the proxy settings
# from the Maven conf/settings.xml file
#
# org.ops4j.pax.url.mvn.proxySupport=false
#
# Comma separated list of repositories scanned when resolving an artifact.
# Those repositories will be checked before iterating through the
# below list of repositories and even before the local repository
# A repository url can be appended with zero or more of the following flags:
# @snapshots : the repository contains snaphots
# @noreleases : the repository does not contain any released artifacts
#
# The following property value will add the system folder as a repo.
#
org.ops4j.pax.url.mvn.defaultRepositories=\
file:${karaf.home}/${karaf.default.repository}@id=system.repository@snapshots,\
file:${karaf.data}/kar@id=kar.repository@multi@snapshots,\
file:${karaf.base}/${karaf.default.repository}@id=child.system.repository@snapshots
# Use the default local repo (e.g.~/.m2/repository) as a "remote" repo
#org.ops4j.pax.url.mvn.defaultLocalRepoAsRemote=false
#
# Comma separated list of repositories scanned when resolving an artifact.
# The default list includes the following repositories:
# http://repo1.maven.org/maven2@id=central
# http://repository.springsource.com/maven/bundles/release@id=spring.ebr
# http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external
# http://zodiac.springsource.com/maven/bundles/release@id=gemini
# http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases
# https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases
# https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases
# To add repositories to the default ones, prepend '+' to the list of repositories
# to add.
# A repository url can be appended with zero or more of the following flags:
# @snapshots : the repository contains snapshots
# @noreleases : the repository does not contain any released artifacts
# @id=repository.id : the id for the repository, just like in the settings.xml this is optional but recommended
#
org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases
### ^^^ No remote repositories. This is the only ODL change compared to Karaf defaults.Configuring the startup features...
+ [[ True == \T\r\u\e ]]
+ echo 'Configuring the startup features...'
+ sed -ie 's/\(featuresBoot=\|featuresBoot =\)/featuresBoot = odl-integration-compatible-with-all,odl-jolokia,odl-restconf,odl-clustering-test-app,/g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
+ FEATURE_TEST_STRING=features-test
+ FEATURE_TEST_VERSION=0.21.3-SNAPSHOT
+ KARAF_VERSION=karaf4
+ [[ integration == \i\n\t\e\g\r\a\t\i\o\n ]]
+ sed -ie 's%\(featuresRepositories=\|featuresRepositories =\)%featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.3-SNAPSHOT/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features,%g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
+ [[ ! -z '' ]]
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
#
# Comma separated list of features repositories to register by default
#
featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.3-SNAPSHOT/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features, file:${karaf.etc}/943186ae-6ff6-48c1-8326-89a825229f44.xml
#
# Comma separated list of features to install at startup
#
featuresBoot = odl-integration-compatible-with-all,odl-jolokia,odl-restconf,odl-clustering-test-app, c6badae2-b1f5-49f1-9c6d-753f26476ccd
#
# Resource repositories (OBR) that the features resolver can use
# to resolve requirements/capabilities
#
# The format of the resourceRepositories is
# resourceRepositories=[xml:url|json:url],...
# for Instance:
#
#resourceRepositories=xml:http://host/path/to/index.xml
# or
#resourceRepositories=json:http://host/path/to/index.json
#
#
# Defines if the boot features are started in asynchronous mode (in a dedicated thread)
#
featuresBootAsynchronous=false
#
# Service requirements enforcement
#
# By default, the feature resolver checks the service requirements/capabilities of
# bundles for new features (xml schema >= 1.3.0) in order to automatically installs
# the required bundles.
# The following flag can have those values:
# - disable: service requirements are completely ignored
# - default: service requirements are ignored for old features
# - enforce: service requirements are always verified
#
#serviceRequirements=default
#
# Store cfg file for config element in feature
#
#configCfgStore=true
#
# Define if the feature service automatically refresh bundles
#
autoRefresh=true
#
# Configuration of features processing mechanism (overrides, blacklisting, modification of features)
# XML file defines instructions related to features processing
# versions.properties may declare properties to resolve placeholders in XML file
# both files are relative to ${karaf.etc}
#
#featureProcessing=org.apache.karaf.features.xml
#featureProcessingVersions=versions.properties
+ configure_karaf_log karaf4 ''
+ local -r karaf_version=karaf4
+ local -r controllerdebugmap=
+ local logapi=log4j
+ grep log4j2 /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n
log4j2.rootLogger.level = INFO
#log4j2.rootLogger.type = asyncRoot
#log4j2.rootLogger.includeLocation = false
log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile
log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi
log4j2.rootLogger.appenderRef.Console.ref = Console
log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter
log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF}
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL
log4j2.logger.spifly.name = org.apache.aries.spifly
log4j2.logger.spifly.level = WARN
log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit
log4j2.logger.audit.level = INFO
log4j2.logger.audit.additivity = false
log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile
# Console appender not used by default (see log4j2.rootLogger.appenderRefs)
log4j2.appender.console.type = Console
log4j2.appender.console.name = Console
log4j2.appender.console.layout.type = PatternLayout
log4j2.appender.console.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.type = RollingRandomAccessFile
log4j2.appender.rolling.name = RollingFile
log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log
log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i
#log4j2.appender.rolling.immediateFlush = false
log4j2.appender.rolling.append = true
log4j2.appender.rolling.layout.type = PatternLayout
log4j2.appender.rolling.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.policies.type = Policies
log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.rolling.policies.size.size = 64MB
log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy
log4j2.appender.rolling.strategy.max = 7
log4j2.appender.audit.type = RollingRandomAccessFile
log4j2.appender.audit.name = AuditRollingFile
log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log
log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i
log4j2.appender.audit.append = true
log4j2.appender.audit.layout.type = PatternLayout
log4j2.appender.audit.layout.pattern = ${log4j2.pattern}
log4j2.appender.audit.policies.type = Policies
log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.audit.policies.size.size = 8MB
log4j2.appender.audit.strategy.type = DefaultRolloverStrategy
log4j2.appender.audit.strategy.max = 7
log4j2.appender.osgi.type = PaxOsgi
log4j2.appender.osgi.name = PaxOsgi
log4j2.appender.osgi.filter = *
#log4j2.logger.aether.name = shaded.org.eclipse.aether
#log4j2.logger.aether.level = TRACE
#log4j2.logger.http-headers.name = shaded.org.apache.http.headers
#log4j2.logger.http-headers.level = DEBUG
#log4j2.logger.maven.name = org.ops4j.pax.url.mvn
#log4j2.logger.maven.level = TRACE
Configuring the karaf log... karaf_version: karaf4, logapi: log4j2
+ logapi=log4j2
+ echo 'Configuring the karaf log... karaf_version: karaf4, logapi: log4j2'
+ '[' log4j2 == log4j2 ']'
+ sed -ie 's/log4j2.appender.rolling.policies.size.size = 64MB/log4j2.appender.rolling.policies.size.size = 1GB/g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
+ orgmodule=org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver
+ orgmodule_=org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver
+ echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN'
+ echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN'
controllerdebugmap:
+ unset IFS
+ echo 'controllerdebugmap: '
+ '[' -n '' ']'
cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
+ echo 'cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
# Common pattern layout for appenders
log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n
# Root logger
log4j2.rootLogger.level = INFO
# uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library
#log4j2.rootLogger.type = asyncRoot
#log4j2.rootLogger.includeLocation = false
log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile
log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi
log4j2.rootLogger.appenderRef.Console.ref = Console
log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter
log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF}
# Filters for logs marked by org.opendaylight.odlparent.Markers
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL
# Loggers configuration
# Spifly logger
log4j2.logger.spifly.name = org.apache.aries.spifly
log4j2.logger.spifly.level = WARN
# Security audit logger
log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit
log4j2.logger.audit.level = INFO
log4j2.logger.audit.additivity = false
log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile
# Appenders configuration
# Console appender not used by default (see log4j2.rootLogger.appenderRefs)
log4j2.appender.console.type = Console
log4j2.appender.console.name = Console
log4j2.appender.console.layout.type = PatternLayout
log4j2.appender.console.layout.pattern = ${log4j2.pattern}
# Rolling file appender
log4j2.appender.rolling.type = RollingRandomAccessFile
log4j2.appender.rolling.name = RollingFile
log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log
log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i
# uncomment to not force a disk flush
#log4j2.appender.rolling.immediateFlush = false
log4j2.appender.rolling.append = true
log4j2.appender.rolling.layout.type = PatternLayout
log4j2.appender.rolling.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.policies.type = Policies
log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.rolling.policies.size.size = 1GB
log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy
log4j2.appender.rolling.strategy.max = 7
# Audit file appender
log4j2.appender.audit.type = RollingRandomAccessFile
log4j2.appender.audit.name = AuditRollingFile
log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log
log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i
log4j2.appender.audit.append = true
log4j2.appender.audit.layout.type = PatternLayout
log4j2.appender.audit.layout.pattern = ${log4j2.pattern}
log4j2.appender.audit.policies.type = Policies
log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.audit.policies.size.size = 8MB
log4j2.appender.audit.strategy.type = DefaultRolloverStrategy
log4j2.appender.audit.strategy.max = 7
# OSGi appender
log4j2.appender.osgi.type = PaxOsgi
log4j2.appender.osgi.name = PaxOsgi
log4j2.appender.osgi.filter = *
# help with identification of maven-related problems with pax-url-aether
#log4j2.logger.aether.name = shaded.org.eclipse.aether
#log4j2.logger.aether.level = TRACE
#log4j2.logger.http-headers.name = shaded.org.apache.http.headers
#log4j2.logger.http-headers.level = DEBUG
#log4j2.logger.maven.name = org.ops4j.pax.url.mvn
#log4j2.logger.maven.level = TRACE
log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN
log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN
+ set_java_vars /usr/lib/jvm/java-21-openjdk-amd64 3072m /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ local -r java_home=/usr/lib/jvm/java-21-openjdk-amd64
+ local -r controllermem=3072m
Configure
java home: /usr/lib/jvm/java-21-openjdk-amd64
max memory: 3072m
memconf: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ local -r memconf=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ echo Configure
+ echo ' java home: /usr/lib/jvm/java-21-openjdk-amd64'
+ echo ' max memory: 3072m'
+ echo ' memconf: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv'
+ sed -ie 's%^# export JAVA_HOME%export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-21-openjdk-amd64}%g' /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ sed -ie 's/JAVA_MAX_MEM="2048m"/JAVA_MAX_MEM=3072m/g' /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ echo 'cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
#!/bin/sh
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
#
# handle specific scripts; the SCRIPT_NAME is exactly the name of the Karaf
# script: client, instance, shell, start, status, stop, karaf
#
# if [ "${KARAF_SCRIPT}" == "SCRIPT_NAME" ]; then
# Actions go here...
# fi
#
# general settings which should be applied for all scripts go here; please keep
# in mind that it is possible that scripts might be executed more than once, e.g.
# in example of the start script where the start script is executed first and the
# karaf script afterwards.
#
#
# The following section shows the possible configuration options for the default
# karaf scripts
#
export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-21-openjdk-amd64} # Location of Java installation
# export JAVA_OPTS # Generic JVM options, for instance, where you can pass the memory configuration
# export JAVA_NON_DEBUG_OPTS # Additional non-debug JVM options
# export EXTRA_JAVA_OPTS # Additional JVM options
# export KARAF_HOME # Karaf home folder
# export KARAF_DATA # Karaf data folder
# export KARAF_BASE # Karaf base folder
# export KARAF_ETC # Karaf etc folder
# export KARAF_LOG # Karaf log folder
# export KARAF_SYSTEM_OPTS # First citizen Karaf options
# export KARAF_OPTS # Additional available Karaf options
# export KARAF_DEBUG # Enable debug mode
# export KARAF_REDIRECT # Enable/set the std/err redirection when using bin/start
# export KARAF_NOROOT # Prevent execution as root if set to true
Set Java version
+ echo 'Set Java version'
+ sudo /usr/sbin/alternatives --install /usr/bin/java java /usr/lib/jvm/java-21-openjdk-amd64/bin/java 1
sudo: a terminal is required to read the password; either use the -S option to read from standard input or configure an askpass helper
sudo: a password is required
+ sudo /usr/sbin/alternatives --set java /usr/lib/jvm/java-21-openjdk-amd64/bin/java
sudo: a terminal is required to read the password; either use the -S option to read from standard input or configure an askpass helper
sudo: a password is required
JDK default version ...
+ echo 'JDK default version ...'
+ java -version
openjdk version "21.0.5" 2024-10-15
OpenJDK Runtime Environment (build 21.0.5+11-Ubuntu-1ubuntu122.04)
OpenJDK 64-Bit Server VM (build 21.0.5+11-Ubuntu-1ubuntu122.04, mixed mode, sharing)
Set JAVA_HOME
+ echo 'Set JAVA_HOME'
+ export JAVA_HOME=/usr/lib/jvm/java-21-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-21-openjdk-amd64
++ readlink -e /usr/lib/jvm/java-21-openjdk-amd64/bin/java
Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-21-openjdk-amd64/bin/java
Listing all open ports on controller system...
+ JAVA_RESOLVED=/usr/lib/jvm/java-21-openjdk-amd64/bin/java
+ echo 'Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-21-openjdk-amd64/bin/java'
+ echo 'Listing all open ports on controller system...'
+ netstat -pnatu
/tmp/configuration-script.sh: line 40: netstat: command not found
+ '[' -f /tmp/custom_shard_config.txt ']'
+ echo 'Custom shard config exists!!!'
Custom shard config exists!!!
Copying the shard config...
+ echo 'Copying the shard config...'
+ cp /tmp/custom_shard_config.txt /tmp/karaf-0.21.3-SNAPSHOT/bin/
Configuring cluster
+ echo 'Configuring cluster'
+ /tmp/karaf-0.21.3-SNAPSHOT/bin/configure_cluster.sh 1 10.30.170.103 10.30.171.180 10.30.171.153
################################################
## Configure Cluster ##
################################################
NOTE: Cluster configuration files not found. Copying from
/tmp/karaf-0.21.3-SNAPSHOT/system/org/opendaylight/controller/sal-clustering-config/10.0.14
Configuring unique name in akka.conf
Configuring hostname in akka.conf
Configuring data and rpc seed nodes in akka.conf
modules = [
{
name = "inventory"
namespace = "urn:opendaylight:inventory"
shard-strategy = "module"
},
{
name = "topology"
namespace = "urn:TBD:params:xml:ns:yang:network-topology"
shard-strategy = "module"
},
{
name = "toaster"
namespace = "http://netconfcentral.org/ns/toaster"
shard-strategy = "module"
},
{
name = "car"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car"
shard-strategy = "module"
},
{
name = "people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:people"
shard-strategy = "module"
},
{
name = "car-people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car-people"
shard-strategy = "module"
}
]
Configuring replication type in module-shards.conf
################################################
## NOTE: Manually restart controller to ##
## apply configuration. ##
################################################
Dump akka.conf
+ echo 'Dump akka.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
odl-cluster-data {
akka {
remote {
artery {
enabled = on
transport = tcp
canonical.hostname = "10.30.170.103"
canonical.port = 2550
}
}
cluster {
# Using artery.
seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.103:2550",
"akka://opendaylight-cluster-data@10.30.171.180:2550",
"akka://opendaylight-cluster-data@10.30.171.153:2550"]
roles = ["member-1"]
# when under load we might trip a false positive on the failure detector
# failure-detector {
# heartbeat-interval = 4 s
# acceptable-heartbeat-pause = 16s
# }
}
persistence {
# By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by
# modifying the following two properties. The directory location specified may be a relative or absolute path.
# The relative path is always relative to KARAF_HOME.
# snapshot-store.local.dir = "target/snapshots"
# Use lz4 compression for LocalSnapshotStore snapshots
snapshot-store.local.use-lz4-compression = false
# Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB
snapshot-store.local.lz4-blocksize = 256KB
}
disable-default-actor-system-quarantined-event-handling = "false"
}
}
Dump modules.conf
+ echo 'Dump modules.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
modules = [
{
name = "inventory"
namespace = "urn:opendaylight:inventory"
shard-strategy = "module"
},
{
name = "topology"
namespace = "urn:TBD:params:xml:ns:yang:network-topology"
shard-strategy = "module"
},
{
name = "toaster"
namespace = "http://netconfcentral.org/ns/toaster"
shard-strategy = "module"
},
{
name = "car"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car"
shard-strategy = "module"
},
{
name = "people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:people"
shard-strategy = "module"
},
{
name = "car-people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car-people"
shard-strategy = "module"
}
]
Dump module-shards.conf
+ echo 'Dump module-shards.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
module-shards = [
{
name = "default"
shards = [
{
name = "default"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "inventory"
shards = [
{
name="inventory"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "topology"
shards = [
{
name="topology"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "toaster"
shards = [
{
name="toaster"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "car"
shards = [
{
name="car"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "people"
shards = [
{
name="people"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "car-people"
shards = [
{
name="car-people"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
}
]
Configuring member-2 with IP address 10.30.171.180
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ source /tmp/common-functions.sh karaf-0.21.3-SNAPSHOT scandium
common-functions.sh is being sourced
++ [[ /tmp/common-functions.sh == \/\t\m\p\/\c\o\n\f\i\g\u\r\a\t\i\o\n\-\s\c\r\i\p\t\.\s\h ]]
++ echo 'common-functions.sh is being sourced'
++ BUNDLEFOLDER=karaf-0.21.3-SNAPSHOT
++ DISTROSTREAM=scandium
++ export MAVENCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
++ MAVENCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
++ export FEATURESCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
++ FEATURESCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
++ export CUSTOMPROP=/tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
++ CUSTOMPROP=/tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
++ export LOGCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
++ LOGCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
++ export MEMCONF=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
++ MEMCONF=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
++ export CONTROLLERMEM=
++ CONTROLLERMEM=
++ case "${DISTROSTREAM}" in
++ CLUSTER_SYSTEM=akka
++ export AKKACONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
++ AKKACONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
++ export MODULESCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
++ MODULESCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
++ export MODULESHARDSCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
++ MODULESHARDSCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
++ print_common_env
++ cat
common-functions environment:
MAVENCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
ACTUALFEATURES:
FEATURESCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
CUSTOMPROP: /tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
LOGCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
MEMCONF: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
CONTROLLERMEM:
AKKACONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
MODULESCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
MODULESHARDSCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
SUITES:
++ SSH='ssh -t -t'
++ extra_services_cntl=' dnsmasq.service httpd.service libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service rabbitmq-server.service '
++ extra_services_cmp=' libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service '
Changing to /tmp
Downloading the distribution from https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
+ echo 'Changing to /tmp'
+ cd /tmp
+ echo 'Downloading the distribution from https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip'
+ wget --progress=dot:mega https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
--2025-10-06 19:31:09-- https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
Resolving nexus.opendaylight.org (nexus.opendaylight.org)... 199.204.45.87, 2604:e100:1:0:f816:3eff:fe45:48d6
Connecting to nexus.opendaylight.org (nexus.opendaylight.org)|199.204.45.87|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 190794064 (182M) [application/zip]
Saving to: ‘karaf-0.21.3-20251006.183506-364.zip’
0K ........ ........ ........ ........ ........ ........ 1% 68.6M 3s
3072K ........ ........ ........ ........ ........ ........ 3% 119M 2s
6144K ........ ........ ........ ........ ........ ........ 4% 150M 2s
9216K ........ ........ ........ ........ ........ ........ 6% 136M 2s
12288K ........ ........ ........ ........ ........ ........ 8% 136M 1s
15360K ........ ........ ........ ........ ........ ........ 9% 178M 1s
18432K ........ ........ ........ ........ ........ ........ 11% 174M 1s
21504K ........ ........ ........ ........ ........ ........ 13% 208M 1s
24576K ........ ........ ........ ........ ........ ........ 14% 225M 1s
27648K ........ ........ ........ ........ ........ ........ 16% 235M 1s
30720K ........ ........ ........ ........ ........ ........ 18% 222M 1s
33792K ........ ........ ........ ........ ........ ........ 19% 251M 1s
36864K ........ ........ ........ ........ ........ ........ 21% 299M 1s
39936K ........ ........ ........ ........ ........ ........ 23% 289M 1s
43008K ........ ........ ........ ........ ........ ........ 24% 274M 1s
46080K ........ ........ ........ ........ ........ ........ 26% 290M 1s
49152K ........ ........ ........ ........ ........ ........ 28% 264M 1s
52224K ........ ........ ........ ........ ........ ........ 29% 261M 1s
55296K ........ ........ ........ ........ ........ ........ 31% 257M 1s
58368K ........ ........ ........ ........ ........ ........ 32% 260M 1s
61440K ........ ........ ........ ........ ........ ........ 34% 252M 1s
64512K ........ ........ ........ ........ ........ ........ 36% 256M 1s
67584K ........ ........ ........ ........ ........ ........ 37% 255M 1s
70656K ........ ........ ........ ........ ........ ........ 39% 253M 1s
73728K ........ ........ ........ ........ ........ ........ 41% 263M 1s
76800K ........ ........ ........ ........ ........ ........ 42% 222M 1s
79872K ........ ........ ........ ........ ........ ........ 44% 251M 1s
82944K ........ ........ ........ ........ ........ ........ 46% 201M 0s
86016K ........ ........ ........ ........ ........ ........ 47% 240M 0s
89088K ........ ........ ........ ........ ........ ........ 49% 242M 0s
92160K ........ ........ ........ ........ ........ ........ 51% 256M 0s
95232K ........ ........ ........ ........ ........ ........ 52% 261M 0s
98304K ........ ........ ........ ........ ........ ........ 54% 258M 0s
101376K ........ ........ ........ ........ ........ ........ 56% 234M 0s
104448K ........ ........ ........ ........ ........ ........ 57% 238M 0s
107520K ........ ........ ........ ........ ........ ........ 59% 248M 0s
110592K ........ ........ ........ ........ ........ ........ 61% 236M 0s
113664K ........ ........ ........ ........ ........ ........ 62% 233M 0s
116736K ........ ........ ........ ........ ........ ........ 64% 234M 0s
119808K ........ ........ ........ ........ ........ ........ 65% 228M 0s
122880K ........ ........ ........ ........ ........ ........ 67% 232M 0s
125952K ........ ........ ........ ........ ........ ........ 69% 227M 0s
129024K ........ ........ ........ ........ ........ ........ 70% 232M 0s
132096K ........ ........ ........ ........ ........ ........ 72% 235M 0s
135168K ........ ........ ........ ........ ........ ........ 74% 235M 0s
138240K ........ ........ ........ ........ ........ ........ 75% 230M 0s
141312K ........ ........ ........ ........ ........ ........ 77% 238M 0s
144384K ........ ........ ........ ........ ........ ........ 79% 230M 0s
147456K ........ ........ ........ ........ ........ ........ 80% 227M 0s
150528K ........ ........ ........ ........ ........ ........ 82% 231M 0s
153600K ........ ........ ........ ........ ........ ........ 84% 224M 0s
156672K ........ ........ ........ ........ ........ ........ 85% 234M 0s
159744K ........ ........ ........ ........ ........ ........ 87% 226M 0s
162816K ........ ........ ........ ........ ........ ........ 89% 228M 0s
165888K ........ ........ ........ ........ ........ ........ 90% 227M 0s
168960K ........ ........ ........ ........ ........ ........ 92% 233M 0s
172032K ........ ........ ........ ........ ........ ........ 93% 225M 0s
175104K ........ ........ ........ ........ ........ ........ 95% 219M 0s
178176K ........ ........ ........ ........ ........ ........ 97% 212M 0s
181248K ........ ........ ........ ........ ........ ........ 98% 220M 0s
184320K ........ ........ ........ ....... 100% 230M=0.8s
2025-10-06 19:31:10 (216 MB/s) - ‘karaf-0.21.3-20251006.183506-364.zip’ saved [190794064/190794064]
Extracting the new controller...
+ echo 'Extracting the new controller...'
+ unzip -q karaf-0.21.3-20251006.183506-364.zip
Adding external repositories...
+ echo 'Adding external repositories...'
+ sed -ie 's%org.ops4j.pax.url.mvn.repositories=%org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases%g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
#
# If set to true, the following property will not allow any certificate to be used
# when accessing Maven repositories through SSL
#
#org.ops4j.pax.url.mvn.certificateCheck=
#
# Path to the local Maven settings file.
# The repositories defined in this file will be automatically added to the list
# of default repositories if the 'org.ops4j.pax.url.mvn.repositories' property
# below is not set.
# The following locations are checked for the existence of the settings.xml file
# * 1. looks for the specified url
# * 2. if not found looks for ${user.home}/.m2/settings.xml
# * 3. if not found looks for ${maven.home}/conf/settings.xml
# * 4. if not found looks for ${M2_HOME}/conf/settings.xml
#
#org.ops4j.pax.url.mvn.settings=
#
# Path to the local Maven repository which is used to avoid downloading
# artifacts when they already exist locally.
# The value of this property will be extracted from the settings.xml file
# above, or defaulted to:
# System.getProperty( "user.home" ) + "/.m2/repository"
#
org.ops4j.pax.url.mvn.localRepository=${karaf.home}/${karaf.default.repository}
#
# Default this to false. It's just weird to use undocumented repos
#
org.ops4j.pax.url.mvn.useFallbackRepositories=false
#
# Uncomment if you don't wanna use the proxy settings
# from the Maven conf/settings.xml file
#
# org.ops4j.pax.url.mvn.proxySupport=false
#
# Comma separated list of repositories scanned when resolving an artifact.
# Those repositories will be checked before iterating through the
# below list of repositories and even before the local repository
# A repository url can be appended with zero or more of the following flags:
# @snapshots : the repository contains snaphots
# @noreleases : the repository does not contain any released artifacts
#
# The following property value will add the system folder as a repo.
#
org.ops4j.pax.url.mvn.defaultRepositories=\
file:${karaf.home}/${karaf.default.repository}@id=system.repository@snapshots,\
file:${karaf.data}/kar@id=kar.repository@multi@snapshots,\
file:${karaf.base}/${karaf.default.repository}@id=child.system.repository@snapshots
# Use the default local repo (e.g.~/.m2/repository) as a "remote" repo
#org.ops4j.pax.url.mvn.defaultLocalRepoAsRemote=false
#
# Comma separated list of repositories scanned when resolving an artifact.
# The default list includes the following repositories:
# http://repo1.maven.org/maven2@id=central
# http://repository.springsource.com/maven/bundles/release@id=spring.ebr
# http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external
# http://zodiac.springsource.com/maven/bundles/release@id=gemini
# http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases
# https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases
# https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases
# To add repositories to the default ones, prepend '+' to the list of repositories
# to add.
# A repository url can be appended with zero or more of the following flags:
# @snapshots : the repository contains snapshots
# @noreleases : the repository does not contain any released artifacts
# @id=repository.id : the id for the repository, just like in the settings.xml this is optional but recommended
#
org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases
### ^^^ No remote repositories. This is the only ODL change compared to Karaf defaults.Configuring the startup features...
+ [[ True == \T\r\u\e ]]
+ echo 'Configuring the startup features...'
+ sed -ie 's/\(featuresBoot=\|featuresBoot =\)/featuresBoot = odl-integration-compatible-with-all,odl-jolokia,odl-restconf,odl-clustering-test-app,/g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
+ FEATURE_TEST_STRING=features-test
+ FEATURE_TEST_VERSION=0.21.3-SNAPSHOT
+ KARAF_VERSION=karaf4
+ [[ integration == \i\n\t\e\g\r\a\t\i\o\n ]]
+ sed -ie 's%\(featuresRepositories=\|featuresRepositories =\)%featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.3-SNAPSHOT/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features,%g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
+ [[ ! -z '' ]]
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
#
# Comma separated list of features repositories to register by default
#
featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.3-SNAPSHOT/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features, file:${karaf.etc}/943186ae-6ff6-48c1-8326-89a825229f44.xml
#
# Comma separated list of features to install at startup
#
featuresBoot = odl-integration-compatible-with-all,odl-jolokia,odl-restconf,odl-clustering-test-app, c6badae2-b1f5-49f1-9c6d-753f26476ccd
#
# Resource repositories (OBR) that the features resolver can use
# to resolve requirements/capabilities
#
# The format of the resourceRepositories is
# resourceRepositories=[xml:url|json:url],...
# for Instance:
#
#resourceRepositories=xml:http://host/path/to/index.xml
# or
#resourceRepositories=json:http://host/path/to/index.json
#
#
# Defines if the boot features are started in asynchronous mode (in a dedicated thread)
#
featuresBootAsynchronous=false
#
# Service requirements enforcement
#
# By default, the feature resolver checks the service requirements/capabilities of
# bundles for new features (xml schema >= 1.3.0) in order to automatically installs
# the required bundles.
# The following flag can have those values:
# - disable: service requirements are completely ignored
# - default: service requirements are ignored for old features
# - enforce: service requirements are always verified
#
#serviceRequirements=default
#
# Store cfg file for config element in feature
#
#configCfgStore=true
#
# Define if the feature service automatically refresh bundles
#
autoRefresh=true
#
# Configuration of features processing mechanism (overrides, blacklisting, modification of features)
# XML file defines instructions related to features processing
# versions.properties may declare properties to resolve placeholders in XML file
# both files are relative to ${karaf.etc}
#
#featureProcessing=org.apache.karaf.features.xml
#featureProcessingVersions=versions.properties
+ configure_karaf_log karaf4 ''
+ local -r karaf_version=karaf4
+ local -r controllerdebugmap=
+ local logapi=log4j
+ grep log4j2 /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n
log4j2.rootLogger.level = INFO
#log4j2.rootLogger.type = asyncRoot
#log4j2.rootLogger.includeLocation = false
log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile
log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi
log4j2.rootLogger.appenderRef.Console.ref = Console
log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter
log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF}
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL
log4j2.logger.spifly.name = org.apache.aries.spifly
log4j2.logger.spifly.level = WARN
log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit
log4j2.logger.audit.level = INFO
log4j2.logger.audit.additivity = false
log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile
# Console appender not used by default (see log4j2.rootLogger.appenderRefs)
log4j2.appender.console.type = Console
log4j2.appender.console.name = Console
log4j2.appender.console.layout.type = PatternLayout
log4j2.appender.console.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.type = RollingRandomAccessFile
log4j2.appender.rolling.name = RollingFile
log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log
log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i
#log4j2.appender.rolling.immediateFlush = false
log4j2.appender.rolling.append = true
log4j2.appender.rolling.layout.type = PatternLayout
log4j2.appender.rolling.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.policies.type = Policies
log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.rolling.policies.size.size = 64MB
log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy
log4j2.appender.rolling.strategy.max = 7
log4j2.appender.audit.type = RollingRandomAccessFile
log4j2.appender.audit.name = AuditRollingFile
log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log
log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i
log4j2.appender.audit.append = true
log4j2.appender.audit.layout.type = PatternLayout
log4j2.appender.audit.layout.pattern = ${log4j2.pattern}
log4j2.appender.audit.policies.type = Policies
log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.audit.policies.size.size = 8MB
log4j2.appender.audit.strategy.type = DefaultRolloverStrategy
log4j2.appender.audit.strategy.max = 7
log4j2.appender.osgi.type = PaxOsgi
log4j2.appender.osgi.name = PaxOsgi
log4j2.appender.osgi.filter = *
#log4j2.logger.aether.name = shaded.org.eclipse.aether
#log4j2.logger.aether.level = TRACE
#log4j2.logger.http-headers.name = shaded.org.apache.http.headers
#log4j2.logger.http-headers.level = DEBUG
#log4j2.logger.maven.name = org.ops4j.pax.url.mvn
#log4j2.logger.maven.level = TRACE
+ logapi=log4j2
+ echo 'Configuring the karaf log... karaf_version: karaf4, logapi: log4j2'
Configuring the karaf log... karaf_version: karaf4, logapi: log4j2
+ '[' log4j2 == log4j2 ']'
+ sed -ie 's/log4j2.appender.rolling.policies.size.size = 64MB/log4j2.appender.rolling.policies.size.size = 1GB/g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
+ orgmodule=org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver
+ orgmodule_=org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver
+ echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN'
controllerdebugmap:
cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
+ echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN'
+ unset IFS
+ echo 'controllerdebugmap: '
+ '[' -n '' ']'
+ echo 'cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
# Common pattern layout for appenders
log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n
# Root logger
log4j2.rootLogger.level = INFO
# uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library
#log4j2.rootLogger.type = asyncRoot
#log4j2.rootLogger.includeLocation = false
log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile
log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi
log4j2.rootLogger.appenderRef.Console.ref = Console
log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter
log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF}
# Filters for logs marked by org.opendaylight.odlparent.Markers
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL
# Loggers configuration
# Spifly logger
log4j2.logger.spifly.name = org.apache.aries.spifly
log4j2.logger.spifly.level = WARN
# Security audit logger
log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit
log4j2.logger.audit.level = INFO
log4j2.logger.audit.additivity = false
log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile
# Appenders configuration
# Console appender not used by default (see log4j2.rootLogger.appenderRefs)
log4j2.appender.console.type = Console
log4j2.appender.console.name = Console
log4j2.appender.console.layout.type = PatternLayout
log4j2.appender.console.layout.pattern = ${log4j2.pattern}
# Rolling file appender
log4j2.appender.rolling.type = RollingRandomAccessFile
log4j2.appender.rolling.name = RollingFile
log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log
log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i
# uncomment to not force a disk flush
#log4j2.appender.rolling.immediateFlush = false
log4j2.appender.rolling.append = true
log4j2.appender.rolling.layout.type = PatternLayout
log4j2.appender.rolling.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.policies.type = Policies
log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.rolling.policies.size.size = 1GB
log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy
log4j2.appender.rolling.strategy.max = 7
# Audit file appender
log4j2.appender.audit.type = RollingRandomAccessFile
log4j2.appender.audit.name = AuditRollingFile
log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log
log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i
log4j2.appender.audit.append = true
log4j2.appender.audit.layout.type = PatternLayout
log4j2.appender.audit.layout.pattern = ${log4j2.pattern}
log4j2.appender.audit.policies.type = Policies
log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.audit.policies.size.size = 8MB
log4j2.appender.audit.strategy.type = DefaultRolloverStrategy
log4j2.appender.audit.strategy.max = 7
# OSGi appender
log4j2.appender.osgi.type = PaxOsgi
log4j2.appender.osgi.name = PaxOsgi
log4j2.appender.osgi.filter = *
# help with identification of maven-related problems with pax-url-aether
#log4j2.logger.aether.name = shaded.org.eclipse.aether
#log4j2.logger.aether.level = TRACE
#log4j2.logger.http-headers.name = shaded.org.apache.http.headers
#log4j2.logger.http-headers.level = DEBUG
#log4j2.logger.maven.name = org.ops4j.pax.url.mvn
#log4j2.logger.maven.level = TRACE
log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN
log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN
+ set_java_vars /usr/lib/jvm/java-21-openjdk-amd64 3072m /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
Configure
java home: /usr/lib/jvm/java-21-openjdk-amd64
max memory: 3072m
memconf: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ local -r java_home=/usr/lib/jvm/java-21-openjdk-amd64
+ local -r controllermem=3072m
+ local -r memconf=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ echo Configure
+ echo ' java home: /usr/lib/jvm/java-21-openjdk-amd64'
+ echo ' max memory: 3072m'
+ echo ' memconf: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv'
+ sed -ie 's%^# export JAVA_HOME%export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-21-openjdk-amd64}%g' /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ sed -ie 's/JAVA_MAX_MEM="2048m"/JAVA_MAX_MEM=3072m/g' /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ echo 'cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
#!/bin/sh
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
#
# handle specific scripts; the SCRIPT_NAME is exactly the name of the Karaf
# script: client, instance, shell, start, status, stop, karaf
#
# if [ "${KARAF_SCRIPT}" == "SCRIPT_NAME" ]; then
# Actions go here...
# fi
#
# general settings which should be applied for all scripts go here; please keep
# in mind that it is possible that scripts might be executed more than once, e.g.
# in example of the start script where the start script is executed first and the
# karaf script afterwards.
#
#
# The following section shows the possible configuration options for the default
# karaf scripts
#
export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-21-openjdk-amd64} # Location of Java installation
# export JAVA_OPTS # Generic JVM options, for instance, where you can pass the memory configuration
# export JAVA_NON_DEBUG_OPTS # Additional non-debug JVM options
# export EXTRA_JAVA_OPTS # Additional JVM options
# export KARAF_HOME # Karaf home folder
# export KARAF_DATA # Karaf data folder
# export KARAF_BASE # Karaf base folder
# export KARAF_ETC # Karaf etc folder
# export KARAF_LOG # Karaf log folder
# export KARAF_SYSTEM_OPTS # First citizen Karaf options
# export KARAF_OPTS # Additional available Karaf options
# export KARAF_DEBUG # Enable debug mode
# export KARAF_REDIRECT # Enable/set the std/err redirection when using bin/start
# export KARAF_NOROOT # Prevent execution as root if set to true
Set Java version
+ echo 'Set Java version'
+ sudo /usr/sbin/alternatives --install /usr/bin/java java /usr/lib/jvm/java-21-openjdk-amd64/bin/java 1
sudo: a terminal is required to read the password; either use the -S option to read from standard input or configure an askpass helper
sudo: a password is required
+ sudo /usr/sbin/alternatives --set java /usr/lib/jvm/java-21-openjdk-amd64/bin/java
sudo: a terminal is required to read the password; either use the -S option to read from standard input or configure an askpass helper
sudo: a password is required
+ echo 'JDK default version ...'
JDK default version ...
+ java -version
openjdk version "21.0.5" 2024-10-15
OpenJDK Runtime Environment (build 21.0.5+11-Ubuntu-1ubuntu122.04)
OpenJDK 64-Bit Server VM (build 21.0.5+11-Ubuntu-1ubuntu122.04, mixed mode, sharing)
Set JAVA_HOME
+ echo 'Set JAVA_HOME'
+ export JAVA_HOME=/usr/lib/jvm/java-21-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-21-openjdk-amd64
++ readlink -e /usr/lib/jvm/java-21-openjdk-amd64/bin/java
Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-21-openjdk-amd64/bin/java
Listing all open ports on controller system...
+ JAVA_RESOLVED=/usr/lib/jvm/java-21-openjdk-amd64/bin/java
+ echo 'Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-21-openjdk-amd64/bin/java'
+ echo 'Listing all open ports on controller system...'
+ netstat -pnatu
/tmp/configuration-script.sh: line 40: netstat: command not found
Custom shard config exists!!!
Copying the shard config...
+ '[' -f /tmp/custom_shard_config.txt ']'
+ echo 'Custom shard config exists!!!'
+ echo 'Copying the shard config...'
+ cp /tmp/custom_shard_config.txt /tmp/karaf-0.21.3-SNAPSHOT/bin/
Configuring cluster
+ echo 'Configuring cluster'
+ /tmp/karaf-0.21.3-SNAPSHOT/bin/configure_cluster.sh 2 10.30.170.103 10.30.171.180 10.30.171.153
################################################
## Configure Cluster ##
################################################
NOTE: Cluster configuration files not found. Copying from
/tmp/karaf-0.21.3-SNAPSHOT/system/org/opendaylight/controller/sal-clustering-config/10.0.14
Configuring unique name in akka.conf
Configuring hostname in akka.conf
Configuring data and rpc seed nodes in akka.conf
modules = [
{
name = "inventory"
namespace = "urn:opendaylight:inventory"
shard-strategy = "module"
},
{
name = "topology"
namespace = "urn:TBD:params:xml:ns:yang:network-topology"
shard-strategy = "module"
},
{
name = "toaster"
namespace = "http://netconfcentral.org/ns/toaster"
shard-strategy = "module"
},
{
name = "car"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car"
shard-strategy = "module"
},
{
name = "people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:people"
shard-strategy = "module"
},
{
name = "car-people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car-people"
shard-strategy = "module"
}
]
Configuring replication type in module-shards.conf
################################################
## NOTE: Manually restart controller to ##
## apply configuration. ##
################################################
Dump akka.conf
+ echo 'Dump akka.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
odl-cluster-data {
akka {
remote {
artery {
enabled = on
transport = tcp
canonical.hostname = "10.30.171.180"
canonical.port = 2550
}
}
cluster {
# Using artery.
seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.103:2550",
"akka://opendaylight-cluster-data@10.30.171.180:2550",
"akka://opendaylight-cluster-data@10.30.171.153:2550"]
roles = ["member-2"]
# when under load we might trip a false positive on the failure detector
# failure-detector {
# heartbeat-interval = 4 s
# acceptable-heartbeat-pause = 16s
# }
}
persistence {
# By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by
# modifying the following two properties. The directory location specified may be a relative or absolute path.
# The relative path is always relative to KARAF_HOME.
# snapshot-store.local.dir = "target/snapshots"
# Use lz4 compression for LocalSnapshotStore snapshots
snapshot-store.local.use-lz4-compression = false
# Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB
snapshot-store.local.lz4-blocksize = 256KB
}
disable-default-actor-system-quarantined-event-handling = "false"
}
}
Dump modules.conf
+ echo 'Dump modules.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
modules = [
{
name = "inventory"
namespace = "urn:opendaylight:inventory"
shard-strategy = "module"
},
{
name = "topology"
namespace = "urn:TBD:params:xml:ns:yang:network-topology"
shard-strategy = "module"
},
{
name = "toaster"
namespace = "http://netconfcentral.org/ns/toaster"
shard-strategy = "module"
},
{
name = "car"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car"
shard-strategy = "module"
},
{
name = "people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:people"
shard-strategy = "module"
},
{
name = "car-people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car-people"
shard-strategy = "module"
}
]
Dump module-shards.conf
+ echo 'Dump module-shards.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
module-shards = [
{
name = "default"
shards = [
{
name = "default"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "inventory"
shards = [
{
name="inventory"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "topology"
shards = [
{
name="topology"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "toaster"
shards = [
{
name="toaster"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "car"
shards = [
{
name="car"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "people"
shards = [
{
name="people"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "car-people"
shards = [
{
name="car-people"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
}
]
Configuring member-3 with IP address 10.30.171.153
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ source /tmp/common-functions.sh karaf-0.21.3-SNAPSHOT scandium
++ [[ /tmp/common-functions.sh == \/\t\m\p\/\c\o\n\f\i\g\u\r\a\t\i\o\n\-\s\c\r\i\p\t\.\s\h ]]
common-functions.sh is being sourced
++ echo 'common-functions.sh is being sourced'
++ BUNDLEFOLDER=karaf-0.21.3-SNAPSHOT
++ DISTROSTREAM=scandium
++ export MAVENCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
++ MAVENCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
++ export FEATURESCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
++ FEATURESCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
++ export CUSTOMPROP=/tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
++ CUSTOMPROP=/tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
++ export LOGCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
++ LOGCONF=/tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
++ export MEMCONF=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
++ MEMCONF=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
++ export CONTROLLERMEM=
++ CONTROLLERMEM=
++ case "${DISTROSTREAM}" in
++ CLUSTER_SYSTEM=akka
++ export AKKACONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
++ AKKACONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
++ export MODULESCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
++ MODULESCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
++ export MODULESHARDSCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
++ MODULESHARDSCONF=/tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
++ print_common_env
++ cat
common-functions environment:
MAVENCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
ACTUALFEATURES:
FEATURESCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
CUSTOMPROP: /tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
LOGCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
MEMCONF: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
CONTROLLERMEM:
AKKACONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
MODULESCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
MODULESHARDSCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
SUITES:
++ SSH='ssh -t -t'
++ extra_services_cntl=' dnsmasq.service httpd.service libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service rabbitmq-server.service '
++ extra_services_cmp=' libvirtd.service openvswitch.service ovs-vswitchd.service ovsdb-server.service '
Changing to /tmp
Downloading the distribution from https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
+ echo 'Changing to /tmp'
+ cd /tmp
+ echo 'Downloading the distribution from https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip'
+ wget --progress=dot:mega https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
--2025-10-06 19:31:13-- https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip
Resolving nexus.opendaylight.org (nexus.opendaylight.org)... 199.204.45.87, 2604:e100:1:0:f816:3eff:fe45:48d6
Connecting to nexus.opendaylight.org (nexus.opendaylight.org)|199.204.45.87|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 190794064 (182M) [application/zip]
Saving to: ‘karaf-0.21.3-20251006.183506-364.zip’
0K ........ ........ ........ ........ ........ ........ 1% 84.5M 2s
3072K ........ ........ ........ ........ ........ ........ 3% 125M 2s
6144K ........ ........ ........ ........ ........ ........ 4% 124M 2s
9216K ........ ........ ........ ........ ........ ........ 6% 99.7M 2s
12288K ........ ........ ........ ........ ........ ........ 8% 163M 1s
15360K ........ ........ ........ ........ ........ ........ 9% 180M 1s
18432K ........ ........ ........ ........ ........ ........ 11% 182M 1s
21504K ........ ........ ........ ........ ........ ........ 13% 220M 1s
24576K ........ ........ ........ ........ ........ ........ 14% 213M 1s
27648K ........ ........ ........ ........ ........ ........ 16% 210M 1s
30720K ........ ........ ........ ........ ........ ........ 18% 240M 1s
33792K ........ ........ ........ ........ ........ ........ 19% 223M 1s
36864K ........ ........ ........ ........ ........ ........ 21% 227M 1s
39936K ........ ........ ........ ........ ........ ........ 23% 159M 1s
43008K ........ ........ ........ ........ ........ ........ 24% 112M 1s
46080K ........ ........ ........ ........ ........ ........ 26% 156M 1s
49152K ........ ........ ........ ........ ........ ........ 28% 238M 1s
52224K ........ ........ ........ ........ ........ ........ 29% 253M 1s
55296K ........ ........ ........ ........ ........ ........ 31% 284M 1s
58368K ........ ........ ........ ........ ........ ........ 32% 246M 1s
61440K ........ ........ ........ ........ ........ ........ 34% 264M 1s
64512K ........ ........ ........ ........ ........ ........ 36% 330M 1s
67584K ........ ........ ........ ........ ........ ........ 37% 333M 1s
70656K ........ ........ ........ ........ ........ ........ 39% 351M 1s
73728K ........ ........ ........ ........ ........ ........ 41% 327M 1s
76800K ........ ........ ........ ........ ........ ........ 42% 253M 1s
79872K ........ ........ ........ ........ ........ ........ 44% 251M 1s
82944K ........ ........ ........ ........ ........ ........ 46% 278M 1s
86016K ........ ........ ........ ........ ........ ........ 47% 173M 0s
89088K ........ ........ ........ ........ ........ ........ 49% 329M 0s
92160K ........ ........ ........ ........ ........ ........ 51% 303M 0s
95232K ........ ........ ........ ........ ........ ........ 52% 334M 0s
98304K ........ ........ ........ ........ ........ ........ 54% 332M 0s
101376K ........ ........ ........ ........ ........ ........ 56% 338M 0s
104448K ........ ........ ........ ........ ........ ........ 57% 328M 0s
107520K ........ ........ ........ ........ ........ ........ 59% 287M 0s
110592K ........ ........ ........ ........ ........ ........ 61% 370M 0s
113664K ........ ........ ........ ........ ........ ........ 62% 359M 0s
116736K ........ ........ ........ ........ ........ ........ 64% 312M 0s
119808K ........ ........ ........ ........ ........ ........ 65% 303M 0s
122880K ........ ........ ........ ........ ........ ........ 67% 327M 0s
125952K ........ ........ ........ ........ ........ ........ 69% 259M 0s
129024K ........ ........ ........ ........ ........ ........ 70% 432M 0s
132096K ........ ........ ........ ........ ........ ........ 72% 410M 0s
135168K ........ ........ ........ ........ ........ ........ 74% 350M 0s
138240K ........ ........ ........ ........ ........ ........ 75% 262M 0s
141312K ........ ........ ........ ........ ........ ........ 77% 217M 0s
144384K ........ ........ ........ ........ ........ ........ 79% 223M 0s
147456K ........ ........ ........ ........ ........ ........ 80% 219M 0s
150528K ........ ........ ........ ........ ........ ........ 82% 235M 0s
153600K ........ ........ ........ ........ ........ ........ 84% 260M 0s
156672K ........ ........ ........ ........ ........ ........ 85% 318M 0s
159744K ........ ........ ........ ........ ........ ........ 87% 345M 0s
162816K ........ ........ ........ ........ ........ ........ 89% 331M 0s
165888K ........ ........ ........ ........ ........ ........ 90% 329M 0s
168960K ........ ........ ........ ........ ........ ........ 92% 324M 0s
172032K ........ ........ ........ ........ ........ ........ 93% 328M 0s
175104K ........ ........ ........ ........ ........ ........ 95% 325M 0s
178176K ........ ........ ........ ........ ........ ........ 97% 325M 0s
181248K ........ ........ ........ ........ ........ ........ 98% 312M 0s
184320K ........ ........ ........ ....... 100% 332M=0.8s
2025-10-06 19:31:14 (237 MB/s) - ‘karaf-0.21.3-20251006.183506-364.zip’ saved [190794064/190794064]
Extracting the new controller...
+ echo 'Extracting the new controller...'
+ unzip -q karaf-0.21.3-20251006.183506-364.zip
Adding external repositories...
+ echo 'Adding external repositories...'
+ sed -ie 's%org.ops4j.pax.url.mvn.repositories=%org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases%g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
#
# If set to true, the following property will not allow any certificate to be used
# when accessing Maven repositories through SSL
#
#org.ops4j.pax.url.mvn.certificateCheck=
#
# Path to the local Maven settings file.
# The repositories defined in this file will be automatically added to the list
# of default repositories if the 'org.ops4j.pax.url.mvn.repositories' property
# below is not set.
# The following locations are checked for the existence of the settings.xml file
# * 1. looks for the specified url
# * 2. if not found looks for ${user.home}/.m2/settings.xml
# * 3. if not found looks for ${maven.home}/conf/settings.xml
# * 4. if not found looks for ${M2_HOME}/conf/settings.xml
#
#org.ops4j.pax.url.mvn.settings=
#
# Path to the local Maven repository which is used to avoid downloading
# artifacts when they already exist locally.
# The value of this property will be extracted from the settings.xml file
# above, or defaulted to:
# System.getProperty( "user.home" ) + "/.m2/repository"
#
org.ops4j.pax.url.mvn.localRepository=${karaf.home}/${karaf.default.repository}
#
# Default this to false. It's just weird to use undocumented repos
#
org.ops4j.pax.url.mvn.useFallbackRepositories=false
#
# Uncomment if you don't wanna use the proxy settings
# from the Maven conf/settings.xml file
#
# org.ops4j.pax.url.mvn.proxySupport=false
#
# Comma separated list of repositories scanned when resolving an artifact.
# Those repositories will be checked before iterating through the
# below list of repositories and even before the local repository
# A repository url can be appended with zero or more of the following flags:
# @snapshots : the repository contains snaphots
# @noreleases : the repository does not contain any released artifacts
#
# The following property value will add the system folder as a repo.
#
org.ops4j.pax.url.mvn.defaultRepositories=\
file:${karaf.home}/${karaf.default.repository}@id=system.repository@snapshots,\
file:${karaf.data}/kar@id=kar.repository@multi@snapshots,\
file:${karaf.base}/${karaf.default.repository}@id=child.system.repository@snapshots
# Use the default local repo (e.g.~/.m2/repository) as a "remote" repo
#org.ops4j.pax.url.mvn.defaultLocalRepoAsRemote=false
#
# Comma separated list of repositories scanned when resolving an artifact.
# The default list includes the following repositories:
# http://repo1.maven.org/maven2@id=central
# http://repository.springsource.com/maven/bundles/release@id=spring.ebr
# http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external
# http://zodiac.springsource.com/maven/bundles/release@id=gemini
# http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases
# https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases
# https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases
# To add repositories to the default ones, prepend '+' to the list of repositories
# to add.
# A repository url can be appended with zero or more of the following flags:
# @snapshots : the repository contains snapshots
# @noreleases : the repository does not contain any released artifacts
# @id=repository.id : the id for the repository, just like in the settings.xml this is optional but recommended
#
org.ops4j.pax.url.mvn.repositories=https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot@id=opendaylight-snapshot@snapshots, https://nexus.opendaylight.org/content/repositories/public@id=opendaylight-mirror, http://repo1.maven.org/maven2@id=central, http://repository.springsource.com/maven/bundles/release@id=spring.ebr.release, http://repository.springsource.com/maven/bundles/external@id=spring.ebr.external, http://zodiac.springsource.com/maven/bundles/release@id=gemini, http://repository.apache.org/content/groups/snapshots-group@id=apache@snapshots@noreleases, https://oss.sonatype.org/content/repositories/snapshots@id=sonatype.snapshots.deploy@snapshots@noreleases, https://oss.sonatype.org/content/repositories/ops4j-snapshots@id=ops4j.sonatype.snapshots.deploy@snapshots@noreleases
### ^^^ No remote repositories. This is the only ODL change compared to Karaf defaults.+ [[ True == \T\r\u\e ]]
+ echo 'Configuring the startup features...'
Configuring the startup features...
+ sed -ie 's/\(featuresBoot=\|featuresBoot =\)/featuresBoot = odl-integration-compatible-with-all,odl-jolokia,odl-restconf,odl-clustering-test-app,/g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
+ FEATURE_TEST_STRING=features-test
+ FEATURE_TEST_VERSION=0.21.3-SNAPSHOT
+ KARAF_VERSION=karaf4
+ [[ integration == \i\n\t\e\g\r\a\t\i\o\n ]]
+ sed -ie 's%\(featuresRepositories=\|featuresRepositories =\)%featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.3-SNAPSHOT/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features,%g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
+ [[ ! -z '' ]]
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
#
# Comma separated list of features repositories to register by default
#
featuresRepositories = mvn:org.opendaylight.integration/features-test/0.21.3-SNAPSHOT/xml/features,mvn:org.apache.karaf.decanter/apache-karaf-decanter/1.2.0/xml/features, file:${karaf.etc}/943186ae-6ff6-48c1-8326-89a825229f44.xml
#
# Comma separated list of features to install at startup
#
featuresBoot = odl-integration-compatible-with-all,odl-jolokia,odl-restconf,odl-clustering-test-app, c6badae2-b1f5-49f1-9c6d-753f26476ccd
#
# Resource repositories (OBR) that the features resolver can use
# to resolve requirements/capabilities
#
# The format of the resourceRepositories is
# resourceRepositories=[xml:url|json:url],...
# for Instance:
#
#resourceRepositories=xml:http://host/path/to/index.xml
# or
#resourceRepositories=json:http://host/path/to/index.json
#
#
# Defines if the boot features are started in asynchronous mode (in a dedicated thread)
#
featuresBootAsynchronous=false
#
# Service requirements enforcement
#
# By default, the feature resolver checks the service requirements/capabilities of
# bundles for new features (xml schema >= 1.3.0) in order to automatically installs
# the required bundles.
# The following flag can have those values:
# - disable: service requirements are completely ignored
# - default: service requirements are ignored for old features
# - enforce: service requirements are always verified
#
#serviceRequirements=default
#
# Store cfg file for config element in feature
#
#configCfgStore=true
#
# Define if the feature service automatically refresh bundles
#
autoRefresh=true
#
# Configuration of features processing mechanism (overrides, blacklisting, modification of features)
# XML file defines instructions related to features processing
# versions.properties may declare properties to resolve placeholders in XML file
# both files are relative to ${karaf.etc}
#
#featureProcessing=org.apache.karaf.features.xml
#featureProcessingVersions=versions.properties
+ configure_karaf_log karaf4 ''
+ local -r karaf_version=karaf4
+ local -r controllerdebugmap=
+ local logapi=log4j
+ grep log4j2 /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n
log4j2.rootLogger.level = INFO
#log4j2.rootLogger.type = asyncRoot
#log4j2.rootLogger.includeLocation = false
log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile
log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi
log4j2.rootLogger.appenderRef.Console.ref = Console
log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter
log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF}
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL
log4j2.logger.spifly.name = org.apache.aries.spifly
log4j2.logger.spifly.level = WARN
log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit
log4j2.logger.audit.level = INFO
log4j2.logger.audit.additivity = false
log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile
# Console appender not used by default (see log4j2.rootLogger.appenderRefs)
log4j2.appender.console.type = Console
log4j2.appender.console.name = Console
log4j2.appender.console.layout.type = PatternLayout
log4j2.appender.console.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.type = RollingRandomAccessFile
log4j2.appender.rolling.name = RollingFile
log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log
log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i
#log4j2.appender.rolling.immediateFlush = false
log4j2.appender.rolling.append = true
log4j2.appender.rolling.layout.type = PatternLayout
log4j2.appender.rolling.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.policies.type = Policies
log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.rolling.policies.size.size = 64MB
log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy
log4j2.appender.rolling.strategy.max = 7
log4j2.appender.audit.type = RollingRandomAccessFile
log4j2.appender.audit.name = AuditRollingFile
log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log
log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i
log4j2.appender.audit.append = true
log4j2.appender.audit.layout.type = PatternLayout
log4j2.appender.audit.layout.pattern = ${log4j2.pattern}
log4j2.appender.audit.policies.type = Policies
log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.audit.policies.size.size = 8MB
log4j2.appender.audit.strategy.type = DefaultRolloverStrategy
log4j2.appender.audit.strategy.max = 7
log4j2.appender.osgi.type = PaxOsgi
log4j2.appender.osgi.name = PaxOsgi
log4j2.appender.osgi.filter = *
#log4j2.logger.aether.name = shaded.org.eclipse.aether
#log4j2.logger.aether.level = TRACE
#log4j2.logger.http-headers.name = shaded.org.apache.http.headers
#log4j2.logger.http-headers.level = DEBUG
#log4j2.logger.maven.name = org.ops4j.pax.url.mvn
#log4j2.logger.maven.level = TRACE
Configuring the karaf log... karaf_version: karaf4, logapi: log4j2
+ logapi=log4j2
+ echo 'Configuring the karaf log... karaf_version: karaf4, logapi: log4j2'
+ '[' log4j2 == log4j2 ']'
+ sed -ie 's/log4j2.appender.rolling.policies.size.size = 64MB/log4j2.appender.rolling.policies.size.size = 1GB/g' /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
+ orgmodule=org.opendaylight.yangtools.yang.parser.repo.YangTextSchemaContextResolver
+ orgmodule_=org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver
+ echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN'
controllerdebugmap:
+ echo 'log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN'
+ unset IFS
+ echo 'controllerdebugmap: '
cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
+ '[' -n '' ']'
+ echo 'cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
################################################################################
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
################################################################################
# Common pattern layout for appenders
log4j2.pattern = %d{ISO8601} | %-5p | %-16t | %-32c{1} | %X{bundle.id} - %X{bundle.name} - %X{bundle.version} | %m%n
# Root logger
log4j2.rootLogger.level = INFO
# uncomment to use asynchronous loggers, which require mvn:com.lmax/disruptor/3.3.2 library
#log4j2.rootLogger.type = asyncRoot
#log4j2.rootLogger.includeLocation = false
log4j2.rootLogger.appenderRef.RollingFile.ref = RollingFile
log4j2.rootLogger.appenderRef.PaxOsgi.ref = PaxOsgi
log4j2.rootLogger.appenderRef.Console.ref = Console
log4j2.rootLogger.appenderRef.Console.filter.threshold.type = ThresholdFilter
log4j2.rootLogger.appenderRef.Console.filter.threshold.level = ${karaf.log.console:-OFF}
# Filters for logs marked by org.opendaylight.odlparent.Markers
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.type = ContextMapFilter
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.type = KeyValuePair
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.key = slf4j.marker
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.pair1.value = CONFIDENTIAL
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.operator = or
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMatch = DENY
log4j2.rootLogger.appenderRef.RollingFile.filter.confidential.onMismatch = NEUTRAL
# Loggers configuration
# Spifly logger
log4j2.logger.spifly.name = org.apache.aries.spifly
log4j2.logger.spifly.level = WARN
# Security audit logger
log4j2.logger.audit.name = org.apache.karaf.jaas.modules.audit
log4j2.logger.audit.level = INFO
log4j2.logger.audit.additivity = false
log4j2.logger.audit.appenderRef.AuditRollingFile.ref = AuditRollingFile
# Appenders configuration
# Console appender not used by default (see log4j2.rootLogger.appenderRefs)
log4j2.appender.console.type = Console
log4j2.appender.console.name = Console
log4j2.appender.console.layout.type = PatternLayout
log4j2.appender.console.layout.pattern = ${log4j2.pattern}
# Rolling file appender
log4j2.appender.rolling.type = RollingRandomAccessFile
log4j2.appender.rolling.name = RollingFile
log4j2.appender.rolling.fileName = ${karaf.data}/log/karaf.log
log4j2.appender.rolling.filePattern = ${karaf.data}/log/karaf.log.%i
# uncomment to not force a disk flush
#log4j2.appender.rolling.immediateFlush = false
log4j2.appender.rolling.append = true
log4j2.appender.rolling.layout.type = PatternLayout
log4j2.appender.rolling.layout.pattern = ${log4j2.pattern}
log4j2.appender.rolling.policies.type = Policies
log4j2.appender.rolling.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.rolling.policies.size.size = 1GB
log4j2.appender.rolling.strategy.type = DefaultRolloverStrategy
log4j2.appender.rolling.strategy.max = 7
# Audit file appender
log4j2.appender.audit.type = RollingRandomAccessFile
log4j2.appender.audit.name = AuditRollingFile
log4j2.appender.audit.fileName = ${karaf.data}/security/audit.log
log4j2.appender.audit.filePattern = ${karaf.data}/security/audit.log.%i
log4j2.appender.audit.append = true
log4j2.appender.audit.layout.type = PatternLayout
log4j2.appender.audit.layout.pattern = ${log4j2.pattern}
log4j2.appender.audit.policies.type = Policies
log4j2.appender.audit.policies.size.type = SizeBasedTriggeringPolicy
log4j2.appender.audit.policies.size.size = 8MB
log4j2.appender.audit.strategy.type = DefaultRolloverStrategy
log4j2.appender.audit.strategy.max = 7
# OSGi appender
log4j2.appender.osgi.type = PaxOsgi
log4j2.appender.osgi.name = PaxOsgi
log4j2.appender.osgi.filter = *
# help with identification of maven-related problems with pax-url-aether
#log4j2.logger.aether.name = shaded.org.eclipse.aether
#log4j2.logger.aether.level = TRACE
#log4j2.logger.http-headers.name = shaded.org.apache.http.headers
#log4j2.logger.http-headers.level = DEBUG
#log4j2.logger.maven.name = org.ops4j.pax.url.mvn
#log4j2.logger.maven.level = TRACE
log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.name = WARN
log4j2.logger.org_opendaylight_yangtools_yang_parser_repo_YangTextSchemaContextResolver.level = WARN
Configure
+ set_java_vars /usr/lib/jvm/java-21-openjdk-amd64 3072m /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ local -r java_home=/usr/lib/jvm/java-21-openjdk-amd64
+ local -r controllermem=3072m
+ local -r memconf=/tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ echo Configure
java home: /usr/lib/jvm/java-21-openjdk-amd64
max memory: 3072m
memconf: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ echo ' java home: /usr/lib/jvm/java-21-openjdk-amd64'
+ echo ' max memory: 3072m'
+ echo ' memconf: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv'
+ sed -ie 's%^# export JAVA_HOME%export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-21-openjdk-amd64}%g' /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ sed -ie 's/JAVA_MAX_MEM="2048m"/JAVA_MAX_MEM=3072m/g' /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
+ echo 'cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
#!/bin/sh
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
#
# handle specific scripts; the SCRIPT_NAME is exactly the name of the Karaf
# script: client, instance, shell, start, status, stop, karaf
#
# if [ "${KARAF_SCRIPT}" == "SCRIPT_NAME" ]; then
# Actions go here...
# fi
#
# general settings which should be applied for all scripts go here; please keep
# in mind that it is possible that scripts might be executed more than once, e.g.
# in example of the start script where the start script is executed first and the
# karaf script afterwards.
#
#
# The following section shows the possible configuration options for the default
# karaf scripts
#
export JAVA_HOME=${JAVA_HOME:-/usr/lib/jvm/java-21-openjdk-amd64} # Location of Java installation
# export JAVA_OPTS # Generic JVM options, for instance, where you can pass the memory configuration
# export JAVA_NON_DEBUG_OPTS # Additional non-debug JVM options
# export EXTRA_JAVA_OPTS # Additional JVM options
# export KARAF_HOME # Karaf home folder
# export KARAF_DATA # Karaf data folder
# export KARAF_BASE # Karaf base folder
# export KARAF_ETC # Karaf etc folder
# export KARAF_LOG # Karaf log folder
# export KARAF_SYSTEM_OPTS # First citizen Karaf options
# export KARAF_OPTS # Additional available Karaf options
# export KARAF_DEBUG # Enable debug mode
# export KARAF_REDIRECT # Enable/set the std/err redirection when using bin/start
# export KARAF_NOROOT # Prevent execution as root if set to true
Set Java version
+ echo 'Set Java version'
+ sudo /usr/sbin/alternatives --install /usr/bin/java java /usr/lib/jvm/java-21-openjdk-amd64/bin/java 1
sudo: a terminal is required to read the password; either use the -S option to read from standard input or configure an askpass helper
sudo: a password is required
+ sudo /usr/sbin/alternatives --set java /usr/lib/jvm/java-21-openjdk-amd64/bin/java
sudo: a terminal is required to read the password; either use the -S option to read from standard input or configure an askpass helper
sudo: a password is required
JDK default version ...
+ echo 'JDK default version ...'
+ java -version
openjdk version "21.0.5" 2024-10-15
OpenJDK Runtime Environment (build 21.0.5+11-Ubuntu-1ubuntu122.04)
OpenJDK 64-Bit Server VM (build 21.0.5+11-Ubuntu-1ubuntu122.04, mixed mode, sharing)
Set JAVA_HOME
+ echo 'Set JAVA_HOME'
+ export JAVA_HOME=/usr/lib/jvm/java-21-openjdk-amd64
+ JAVA_HOME=/usr/lib/jvm/java-21-openjdk-amd64
++ readlink -e /usr/lib/jvm/java-21-openjdk-amd64/bin/java
Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-21-openjdk-amd64/bin/java
+ JAVA_RESOLVED=/usr/lib/jvm/java-21-openjdk-amd64/bin/java
+ echo 'Java binary pointed at by JAVA_HOME: /usr/lib/jvm/java-21-openjdk-amd64/bin/java'
Listing all open ports on controller system...
+ echo 'Listing all open ports on controller system...'
+ netstat -pnatu
/tmp/configuration-script.sh: line 40: netstat: command not found
Custom shard config exists!!!
Copying the shard config...
+ '[' -f /tmp/custom_shard_config.txt ']'
+ echo 'Custom shard config exists!!!'
+ echo 'Copying the shard config...'
+ cp /tmp/custom_shard_config.txt /tmp/karaf-0.21.3-SNAPSHOT/bin/
Configuring cluster
+ echo 'Configuring cluster'
+ /tmp/karaf-0.21.3-SNAPSHOT/bin/configure_cluster.sh 3 10.30.170.103 10.30.171.180 10.30.171.153
################################################
## Configure Cluster ##
################################################
NOTE: Cluster configuration files not found. Copying from
/tmp/karaf-0.21.3-SNAPSHOT/system/org/opendaylight/controller/sal-clustering-config/10.0.14
Configuring unique name in akka.conf
Configuring hostname in akka.conf
Configuring data and rpc seed nodes in akka.conf
modules = [
{
name = "inventory"
namespace = "urn:opendaylight:inventory"
shard-strategy = "module"
},
{
name = "topology"
namespace = "urn:TBD:params:xml:ns:yang:network-topology"
shard-strategy = "module"
},
{
name = "toaster"
namespace = "http://netconfcentral.org/ns/toaster"
shard-strategy = "module"
},
{
name = "car"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car"
shard-strategy = "module"
},
{
name = "people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:people"
shard-strategy = "module"
},
{
name = "car-people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car-people"
shard-strategy = "module"
}
]
Configuring replication type in module-shards.conf
################################################
## NOTE: Manually restart controller to ##
## apply configuration. ##
################################################
Dump akka.conf
+ echo 'Dump akka.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
odl-cluster-data {
akka {
remote {
artery {
enabled = on
transport = tcp
canonical.hostname = "10.30.171.153"
canonical.port = 2550
}
}
cluster {
# Using artery.
seed-nodes = ["akka://opendaylight-cluster-data@10.30.170.103:2550",
"akka://opendaylight-cluster-data@10.30.171.180:2550",
"akka://opendaylight-cluster-data@10.30.171.153:2550"]
roles = ["member-3"]
# when under load we might trip a false positive on the failure detector
# failure-detector {
# heartbeat-interval = 4 s
# acceptable-heartbeat-pause = 16s
# }
}
persistence {
# By default the snapshots/journal directories live in KARAF_HOME. You can choose to put it somewhere else by
# modifying the following two properties. The directory location specified may be a relative or absolute path.
# The relative path is always relative to KARAF_HOME.
# snapshot-store.local.dir = "target/snapshots"
# Use lz4 compression for LocalSnapshotStore snapshots
snapshot-store.local.use-lz4-compression = false
# Size of blocks for lz4 compression: 64KB, 256KB, 1MB or 4MB
snapshot-store.local.lz4-blocksize = 256KB
}
disable-default-actor-system-quarantined-event-handling = "false"
}
}
Dump modules.conf
+ echo 'Dump modules.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
modules = [
{
name = "inventory"
namespace = "urn:opendaylight:inventory"
shard-strategy = "module"
},
{
name = "topology"
namespace = "urn:TBD:params:xml:ns:yang:network-topology"
shard-strategy = "module"
},
{
name = "toaster"
namespace = "http://netconfcentral.org/ns/toaster"
shard-strategy = "module"
},
{
name = "car"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car"
shard-strategy = "module"
},
{
name = "people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:people"
shard-strategy = "module"
},
{
name = "car-people"
namespace = "urn:opendaylight:params:xml:ns:yang:controller:config:sal-clustering-it:car-people"
shard-strategy = "module"
}
]
Dump module-shards.conf
+ echo 'Dump module-shards.conf'
+ cat /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
module-shards = [
{
name = "default"
shards = [
{
name = "default"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "inventory"
shards = [
{
name="inventory"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "topology"
shards = [
{
name="topology"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "toaster"
shards = [
{
name="toaster"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "car"
shards = [
{
name="car"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "people"
shards = [
{
name="people"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
},
{
name = "car-people"
shards = [
{
name="car-people"
replicas = ["member-1",
"member-2",
"member-3"]
}
]
}
]
Locating config plan to use...
Finished running config plans
Starting member-1 with IP address 10.30.170.103
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
Redirecting karaf console output to karaf_console.log
Starting controller...
start: Redirecting Karaf output to /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
Starting member-2 with IP address 10.30.171.180
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
Redirecting karaf console output to karaf_console.log
Starting controller...
start: Redirecting Karaf output to /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
Starting member-3 with IP address 10.30.171.153
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
Redirecting karaf console output to karaf_console.log
Starting controller...
start: Redirecting Karaf output to /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins12536295351992642192.sh
common-functions.sh is being sourced
common-functions environment:
MAVENCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.url.mvn.cfg
ACTUALFEATURES:
FEATURESCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.apache.karaf.features.cfg
CUSTOMPROP: /tmp/karaf-0.21.3-SNAPSHOT/etc/custom.properties
LOGCONF: /tmp/karaf-0.21.3-SNAPSHOT/etc/org.ops4j.pax.logging.cfg
MEMCONF: /tmp/karaf-0.21.3-SNAPSHOT/bin/setenv
CONTROLLERMEM: 2048m
AKKACONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/akka.conf
MODULESCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/modules.conf
MODULESHARDSCONF: /tmp/karaf-0.21.3-SNAPSHOT/configuration/initial/module-shards.conf
SUITES:
+ echo '#################################################'
#################################################
+ echo '## Verify Cluster is UP ##'
## Verify Cluster is UP ##
+ echo '#################################################'
#################################################
+ create_post_startup_script
+ cat
+ copy_and_run_post_startup_script
+ seed_index=1
++ seq 1 3
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_1_IP
+ echo 'Execute the post startup script on controller 10.30.170.103'
Execute the post startup script on controller 10.30.170.103
+ scp /w/workspace/controller-csit-3node-clustering-tell-all-scandium/post-startup-script.sh 10.30.170.103:/tmp/
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ ssh 10.30.170.103 'bash /tmp/post-startup-script.sh 1'
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
Waiting up to 3 minutes for controller to come up, checking every 5 seconds...
2025-10-06T19:32:01,696 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 282 - org.opendaylight.infrautils.ready-api - 7.1.7 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.]
Controller is UP
2025-10-06T19:32:01,696 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 282 - org.opendaylight.infrautils.ready-api - 7.1.7 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.]
Listing all open ports on controller system...
/tmp/post-startup-script.sh: line 51: netstat: command not found
looking for "BindException: Address already in use" in log file
looking for "server is unhealthy" in log file
+ '[' 1 == 0 ']'
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_2_IP
+ echo 'Execute the post startup script on controller 10.30.171.180'
Execute the post startup script on controller 10.30.171.180
+ scp /w/workspace/controller-csit-3node-clustering-tell-all-scandium/post-startup-script.sh 10.30.171.180:/tmp/
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ ssh 10.30.171.180 'bash /tmp/post-startup-script.sh 2'
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
Waiting up to 3 minutes for controller to come up, checking every 5 seconds...
2025-10-06T19:32:02,188 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 282 - org.opendaylight.infrautils.ready-api - 7.1.7 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.]
Controller is UP
2025-10-06T19:32:02,188 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 282 - org.opendaylight.infrautils.ready-api - 7.1.7 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.]
Listing all open ports on controller system...
/tmp/post-startup-script.sh: line 51: netstat: command not found
looking for "BindException: Address already in use" in log file
looking for "server is unhealthy" in log file
+ '[' 2 == 0 ']'
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_3_IP
+ echo 'Execute the post startup script on controller 10.30.171.153'
Execute the post startup script on controller 10.30.171.153
+ scp /w/workspace/controller-csit-3node-clustering-tell-all-scandium/post-startup-script.sh 10.30.171.153:/tmp/
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ ssh 10.30.171.153 'bash /tmp/post-startup-script.sh 3'
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
/tmp/post-startup-script.sh: line 4: netstat: command not found
Waiting up to 3 minutes for controller to come up, checking every 5 seconds...
2025-10-06T19:32:04,328 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 282 - org.opendaylight.infrautils.ready-api - 7.1.7 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.]
Controller is UP
2025-10-06T19:32:04,328 | INFO | SystemReadyService-0 | SimpleSystemReadyMonitor | 282 - org.opendaylight.infrautils.ready-api - 7.1.7 | System ready; AKA: Aye captain, all warp coils are now operating at peak efficiency! [M.]
Listing all open ports on controller system...
/tmp/post-startup-script.sh: line 51: netstat: command not found
looking for "BindException: Address already in use" in log file
looking for "server is unhealthy" in log file
+ '[' 0 == 0 ']'
+ seed_index=1
+ dump_controller_threads
++ seq 1 3
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_1_IP
+ echo 'Let'\''s take the karaf thread dump'
Let's take the karaf thread dump
+ ssh 10.30.170.103 'sudo ps aux'
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
++ grep org.apache.karaf.main.Main /w/workspace/controller-csit-3node-clustering-tell-all-scandium/ps_before.log
++ grep -v grep
++ tr -s ' '
++ cut -f2 '-d '
+ pid=2085
+ echo 'karaf main: org.apache.karaf.main.Main, pid:2085'
karaf main: org.apache.karaf.main.Main, pid:2085
+ ssh 10.30.170.103 '/usr/lib/jvm/java-21-openjdk-amd64/bin/jstack -l 2085'
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_2_IP
+ echo 'Let'\''s take the karaf thread dump'
Let's take the karaf thread dump
+ ssh 10.30.171.180 'sudo ps aux'
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
++ grep org.apache.karaf.main.Main /w/workspace/controller-csit-3node-clustering-tell-all-scandium/ps_before.log
++ grep -v grep
++ tr -s ' '
++ cut -f2 '-d '
+ pid=2080
+ echo 'karaf main: org.apache.karaf.main.Main, pid:2080'
karaf main: org.apache.karaf.main.Main, pid:2080
+ ssh 10.30.171.180 '/usr/lib/jvm/java-21-openjdk-amd64/bin/jstack -l 2080'
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_3_IP
+ echo 'Let'\''s take the karaf thread dump'
Let's take the karaf thread dump
+ ssh 10.30.171.153 'sudo ps aux'
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
++ grep org.apache.karaf.main.Main /w/workspace/controller-csit-3node-clustering-tell-all-scandium/ps_before.log
++ grep -v grep
++ tr -s ' '
++ cut -f2 '-d '
+ pid=2082
+ echo 'karaf main: org.apache.karaf.main.Main, pid:2082'
karaf main: org.apache.karaf.main.Main, pid:2082
+ ssh 10.30.171.153 '/usr/lib/jvm/java-21-openjdk-amd64/bin/jstack -l 2082'
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ '[' 0 -gt 0 ']'
+ echo 'Generating controller variables...'
Generating controller variables...
++ seq 1 3
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_1_IP
+ odl_variables=' -v ODL_SYSTEM_1_IP:10.30.170.103'
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_2_IP
+ odl_variables=' -v ODL_SYSTEM_1_IP:10.30.170.103 -v ODL_SYSTEM_2_IP:10.30.171.180'
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_3_IP
+ odl_variables=' -v ODL_SYSTEM_1_IP:10.30.170.103 -v ODL_SYSTEM_2_IP:10.30.171.180 -v ODL_SYSTEM_3_IP:10.30.171.153'
+ echo 'Generating mininet variables...'
Generating mininet variables...
++ seq 1 0
+ get_test_suites SUITES
+ local __suite_list=SUITES
+ echo 'Locating test plan to use...'
Locating test plan to use...
+ testplan_filepath=/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/testplans/controller-clustering-tell-scandium.txt
+ '[' '!' -f /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/testplans/controller-clustering-tell-scandium.txt ']'
+ testplan_filepath=/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/testplans/controller-clustering-tell.txt
+ '[' disabled '!=' disabled ']'
+ echo 'Changing the testplan path...'
Changing the testplan path...
+ sed s:integration:/w/workspace/controller-csit-3node-clustering-tell-all-scandium: /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/testplans/controller-clustering-tell.txt
+ cat testplan.txt
# Place the suites in run order:
# disabling all "prefix based" suites. This functionality was newly
# introduced with these tests in 2017. The assumption is that
# development stopped in that area, but may continue at a later
# time. Still, the tests are failing and no resources to solve the
# failures at this time. If/When this changes, we can uncomment
# those suites. Note that every suite is commented out with it's
# accompanying "restart_with_tell_based_true" setup suite.
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/ddb-sanity-prefix-based.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/ddb-sanity-module-based.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/clean_leader_shutdown.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/clean_leader_shutdown_prefbasedshard.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/explicit_leader_movement.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/explicit_leader_movement_prefbasedshard.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/leader_isolation.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/leader_isolation_prefbasedshard.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/client_isolation.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/client_isolation_prefbasedshard.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/listener_stability.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
# /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/listener_stability_prefbasedshard.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/master_stability.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/partition_and_heal.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/chasing_the_leader.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/carpeople_crud.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_persistence_recovery.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/entity_isolate.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover_isolation.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud_isolation.robot
/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_outage_corners.robot
+ '[' -z '' ']'
++ grep -E -v '(^[[:space:]]*#|^[[:space:]]*$)' testplan.txt
++ tr '\012' ' '
+ suite_list='/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/ddb-sanity-module-based.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/clean_leader_shutdown.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/explicit_leader_movement.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/leader_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/client_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/listener_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/master_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/partition_and_heal.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/chasing_the_leader.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/carpeople_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_persistence_recovery.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/entity_isolate.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_outage_corners.robot '
+ eval 'SUITES='\''/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/ddb-sanity-module-based.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/clean_leader_shutdown.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/explicit_leader_movement.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/leader_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/client_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/listener_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/master_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/partition_and_heal.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/chasing_the_leader.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/carpeople_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_persistence_recovery.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/entity_isolate.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_outage_corners.robot '\'''
++ SUITES='/w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/ddb-sanity-module-based.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/clean_leader_shutdown.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/explicit_leader_movement.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/leader_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/client_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/listener_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/master_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/partition_and_heal.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/chasing_the_leader.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/carpeople_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_persistence_recovery.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/entity_isolate.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_outage_corners.robot '
+ echo 'Starting Robot test suites /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/ddb-sanity-module-based.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/clean_leader_shutdown.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/explicit_leader_movement.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/leader_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/client_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/listener_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/master_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/partition_and_heal.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/chasing_the_leader.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/carpeople_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_persistence_recovery.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/entity_isolate.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_outage_corners.robot ...'
Starting Robot test suites /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/ddb-sanity-module-based.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/clean_leader_shutdown.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/explicit_leader_movement.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/leader_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/client_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/listener_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/master_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/partition_and_heal.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/chasing_the_leader.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/carpeople_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_persistence_recovery.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/entity_isolate.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_outage_corners.robot ...
+ robot -N controller-clustering-tell.txt --removekeywords wuks -e exclude -e skip_if_scandium -v BUNDLEFOLDER:karaf-0.21.3-SNAPSHOT -v BUNDLE_URL:https://nexus.opendaylight.org/content/repositories/opendaylight.snapshot/org/opendaylight/integration/karaf/0.21.3-SNAPSHOT/karaf-0.21.3-20251006.183506-364.zip -v CONTROLLER:10.30.170.103 -v CONTROLLER1:10.30.171.180 -v CONTROLLER2:10.30.171.153 -v CONTROLLER_USER:jenkins -v JAVA_HOME:/usr/lib/jvm/java-21-openjdk-amd64 -v JDKVERSION:openjdk21 -v JENKINS_WORKSPACE:/w/workspace/controller-csit-3node-clustering-tell-all-scandium -v MININET: -v MININET1: -v MININET2: -v MININET_USER:jenkins -v NEXUSURL_PREFIX:https://nexus.opendaylight.org -v NUM_ODL_SYSTEM:3 -v NUM_TOOLS_SYSTEM:0 -v ODL_STREAM:scandium -v ODL_SYSTEM_IP:10.30.170.103 -v ODL_SYSTEM_1_IP:10.30.170.103 -v ODL_SYSTEM_2_IP:10.30.171.180 -v ODL_SYSTEM_3_IP:10.30.171.153 -v ODL_SYSTEM_USER:jenkins -v TOOLS_SYSTEM_IP: -v TOOLS_SYSTEM_USER:jenkins -v USER_HOME:/home/jenkins -v IS_KARAF_APPL:True -v WORKSPACE:/tmp /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/ddb-sanity-module-based.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/clean_leader_shutdown.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/explicit_leader_movement.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/leader_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/client_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/listener_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/master_stability.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/partition_and_heal.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/cluster_singleton/chasing_the_leader.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/dom_data_broker/restart_odl_with_tell_based_true.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/carpeople_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_persistence_recovery.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/entity_isolate.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/buycar_failover_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_failover_crud_isolation.robot /w/workspace/controller-csit-3node-clustering-tell-all-scandium/test/csit/suites/controller/Clustering_Datastore/car_outage_corners.robot
==============================================================================
controller-clustering-tell.txt
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Ddb-Sanity-Module-Based :: DOMDataBroker tes...
==============================================================================
Get_Shard_Role :: Get shard role. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_get_shard_role"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_get_shard_role&order=bug_status"
HTTPError: 404 Client Error: Not Found for url: http://10.30.170.103:8181/restconf/operations/cluster-admin:get-shard-role
------------------------------------------------------------------------------
Subscribe_Listener_To_Leader :: Subscribe listener to leader. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_subscribe_listener_to_leader"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_subscribe_listener_to_leader&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Subscribe_Listener_To_Follower :: Subscribe listener to follower. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_subscribe_listener_to_follower"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_subscribe_listener_to_follower&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Make_Leader_Local :: Make the loeader local and verify. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_make_leader_local"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_make_leader_local&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Remove_Leader_Shard_Replica_And_Add_It_Back :: Remove and add shar... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_remove_leader_shard_replica_and_add_it_back"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_remove_leader_shard_replica_and_add_it_back&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Remove_Follower_Shard_Replica_And_Add_It_Back :: Remove and add sh... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_remove_follower_shard_replica_and_add_it_back"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_remove_follower_shard_replica_and_add_it_back&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Write_Transactions_One_Node_Leader :: Write transactions. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_write_transactions_one_node_leader"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_write_transactions_one_node_leader&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Write_Transactions_One_Node_Follower :: Write transactions. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_write_transactions_one_node_follower"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_write_transactions_one_node_follower&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Write_Transactions_All_Nodes :: Write transactions. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_ddb_sanity_module_based_write_transactions_all_nodes"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_ddb_sanity_module_based_write_transactions_all_nodes&order=bug_status"
Resolving variable '${response.text}' failed: AttributeError: 'ConnectionError' object has no attribute 'text'
------------------------------------------------------------------------------
controller-clustering-tell.txt.Ddb-Sanity-Module-Based :: DOMDataB... | FAIL |
9 tests, 0 passed, 9 failed
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Clean Leader Shutdown :: DOMDataBroker testi...
==============================================================================
Local_Leader_Shutdown :: Shutdown the leader on the same node as t... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_clean_leader_shutdown_local_leader_shutdown"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_clean_leader_shutdown_local_leader_shutdown&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Restart :: Restart odl | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_clean_leader_shutdown_restart"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_clean_leader_shutdown_restart&order=bug_status"
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Remote_Leader_Shutdown :: Shutdown the leader on different node as... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_clean_leader_shutdown_remote_leader_shutdown"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_clean_leader_shutdown_remote_leader_shutdown&order=bug_status"
ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
------------------------------------------------------------------------------
controller-clustering-tell.txt.Clean Leader Shutdown :: DOMDataBro... | FAIL |
3 tests, 0 passed, 3 failed
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Explicit Leader Movement :: DOMDataBroker te...
==============================================================================
Local_To_Remote_Movement :: Leader moves from local to remote node... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_explicit_leader_movement_local_to_remote_movement"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_explicit_leader_movement_local_to_remote_movement&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Restart1 :: Restart odl | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_explicit_leader_movement_restart1"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_explicit_leader_movement_restart1&order=bug_status"
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Remote_To_Remote_Movement :: Leader moves from one remote to other... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_explicit_leader_movement_remote_to_remote_movement"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_explicit_leader_movement_remote_to_remote_movement&order=bug_status"
ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
------------------------------------------------------------------------------
Restart2 :: Restart odl | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_explicit_leader_movement_restart2"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_explicit_leader_movement_restart2&order=bug_status"
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Remote_To_Local_Movement :: Leader moves from remote to local node... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_explicit_leader_movement_remote_to_local_movement"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_explicit_leader_movement_remote_to_local_movement&order=bug_status"
ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
------------------------------------------------------------------------------
controller-clustering-tell.txt.Explicit Leader Movement :: DOMData... | FAIL |
5 tests, 0 passed, 5 failed
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Leader Isolation :: DOMDataBroker testing: L...
==============================================================================
Healing_Within_Request_Timeout :: The isolated node (leader) is re... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_leader_isolation_healing_within_request_timeout"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_leader_isolation_healing_within_request_timeout&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Restart :: Restart odl | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_leader_isolation_restart"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_leader_isolation_restart&order=bug_status"
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Healing_After_Request_Timeout :: The isolated node (leader) is rej... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_leader_isolation_healing_after_request_timeout"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_leader_isolation_healing_after_request_timeout&order=bug_status"
ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
------------------------------------------------------------------------------
controller-clustering-tell.txt.Leader Isolation :: DOMDataBroker t... | FAIL |
3 tests, 0 passed, 3 failed
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Client Isolation :: DOMDataBroker testing: C...
==============================================================================
Producer_On_Shard_Leader_Node_ChainedTx :: Client isolation with p... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_client_isolation_producer_on_shard_leader_node_chainedtx"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_client_isolation_producer_on_shard_leader_node_chainedtx&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
Also keyword teardown failed:
Several failures occurred:
1) Variable '${client_node_dst}' not found.
2) Variable '${client_node_dst}' not found.
3) Keyword 'ShardStability.Shards_Stability_Get_Details' failed after retrying for 1 minute 10 seconds. The last error was: Several failures occurred:
1) ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
2) Resolving variable '${response.text}' failed: AttributeError: 'NoneType' object has no attribute 'text'
3) Resolving variable '${response.status_code}' failed: AttributeError: 'No...
[ Message content over the limit has been removed. ]
119) Resolving variable '${response.text}' failed: AttributeError: 'NoneType' object has no attribute 'text'
120) Evaluating expression "json.loads('''None''')" failed: JSONDecodeError: Expecting value: line 1 column 1 (char 0)
121) TypeError: Expected argument 1 to be a dictionary, got None instead.
122) TypeError: Expected argument 1 to be a dictionary, got None instead.
123) Unrecognized Raft state: None
124) No leader found.
125) Too many Leaders.
126) IndexError: Given index 0 is out of the range 0--1.
4) KeyError: 'threads'
------------------------------------------------------------------------------
Restart1 :: Restart odl | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_client_isolation_restart1"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_client_isolation_restart1&order=bug_status"
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Producer_On_Shard_Leader_Node_SimpleTx :: Client isolation with pr... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_client_isolation_producer_on_shard_leader_node_simpletx"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_client_isolation_producer_on_shard_leader_node_simpletx&order=bug_status"
ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
Also keyword teardown failed:
Several failures occurred:
1) Variable '${client_node_dst}' not found.
2) Variable '${client_node_dst}' not found.
3) Keyword 'ShardStability.Shards_Stability_Get_Details' failed after retrying for 1 minute 10 seconds. The last error was: Several failures occurred:
1) ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
Also keyword teardown failed:
Several failures occurred:
1) Variable '${client_node_dst}' not found.
2) Variable '${client_node_dst}' not found.
3) Keyword 'ShardStability.Shards_Stability_Get_Details' failed after retrying for 1 minute 10 seconds. The last error was: Several failures occurred:
1) ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
Also keyword teardown failed:
Several failures occurred:
1) Variable '${client_node_dst}' not found.
2) Variable '${client_node_dst}' not found.
3) Keyword 'ShardStability.Shards_Stability_Get_Details' failed after retrying for 1 minute 10 seconds. The last error was: Several failures occurred:
1) ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
------------------------------------------------------------------------------
Restart_2 :: Restart odl. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_listener_stability_restart_2"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_listener_stability_restart_2&order=bug_status"
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Move_Leader_From_Listener_Remote_To_Local :: Listener runs on foll... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_listener_stability_move_leader_from_listener_remote_to_local"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_listener_stability_move_leader_from_listener_remote_to_local&order=bug_status"
ConnectionError: HTTPConnectionPool(host='10.30.170.103', port=8181): Max retries exceeded with url: /jolokia/read/org.opendaylight.controller:Category=Shards,name=member-1-shard-default-config,type=DistributedConfigDatastore (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused'))
------------------------------------------------------------------------------
controller-clustering-tell.txt.Listener Stability :: DOMDataBroker... | FAIL |
5 tests, 0 passed, 5 failed
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Master Stability :: Cluster Singleton testin...
==============================================================================
Register_Singleton_Constant_On_Each_Node_And_Verify :: Register a ... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_master_stability_register_singleton_constant_on_each_node_and_verify"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_master_stability_register_singleton_constant_on_each_node_and_verify&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Unregister_Singleton_Constant_On_Non_Master_Node :: Unregister the... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_master_stability_unregister_singleton_constant_on_non_master_node"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_master_stability_unregister_singleton_constant_on_non_master_node&order=bug_status"
Variable '@{cs_candidates}' not found. Did you mean:
@{cs_exp_candidates}
------------------------------------------------------------------------------
Monitor_Stability_While_Unregistered :: Verify that the owner rema... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_master_stability_monitor_stability_while_unregistered"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_master_stability_monitor_stability_while_unregistered&order=bug_status"
Variable '${cs_owner}' not found.
------------------------------------------------------------------------------
Reregister_Singleton_Constant :: Re-register the unregistered cand... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_master_stability_reregister_singleton_constant"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_master_stability_reregister_singleton_constant&order=bug_status"
Variable '${unregistered_node}' not found.
------------------------------------------------------------------------------
Verify_Stability_After_Reregistration :: Verify that the owner rem... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_master_stability_verify_stability_after_reregistration"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_master_stability_verify_stability_after_reregistration&order=bug_status"
Variable '${cs_owner}' not found.
------------------------------------------------------------------------------
Unregister_Singleton_Constant_On_Each_Node :: Unregister the appli... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_master_stability_unregister_singleton_constant_on_each_node"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_master_stability_unregister_singleton_constant_on_each_node&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Master Stability :: Cluster Singlet... | FAIL |
6 tests, 0 passed, 6 failed
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Partition And Heal :: Cluster Singleton test...
==============================================================================
Register_Singleton_Constant_On_Each_Node :: Register a candidate a... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_partition_and_heal_register_singleton_constant_on_each_node"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_partition_and_heal_register_singleton_constant_on_each_node&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Verify_Singleton_Constant_On_Each_Node :: Store the owner and cand... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_partition_and_heal_verify_singleton_constant_on_each_node"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_partition_and_heal_verify_singleton_constant_on_each_node&order=bug_status"
Could not parse owner and candidates for device get-singleton-constant-service']
------------------------------------------------------------------------------
Isolate_Owner_Node :: Isolate the cluster node which is the owner.... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_partition_and_heal_isolate_owner_node"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_partition_and_heal_isolate_owner_node&order=bug_status"
Variable '${cs_owner}' not found.
------------------------------------------------------------------------------
Monitor_Stability_While_Isolated :: Monitor the stability of the s... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_partition_and_heal_monitor_stability_while_isolated"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_partition_and_heal_monitor_stability_while_isolated&order=bug_status"
Invalid IF condition: Evaluating expression '"${index}" == "${cs_isolated_index}"' failed: Variable '${cs_isolated_index}' not found.
------------------------------------------------------------------------------
Rejoin_Isolated_node :: Rejoin isolated node. | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_partition_and_heal_rejoin_isolated_node"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_partition_and_heal_rejoin_isolated_node&order=bug_status"
Variable '${cs_isolated_index}' not found.
------------------------------------------------------------------------------
Unregister_Singleton_Constant_On_Each_Node :: Unregister the appli... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_partition_and_heal_unregister_singleton_constant_on_each_node"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_partition_and_heal_unregister_singleton_constant_on_each_node&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Partition And Heal :: Cluster Singl... | FAIL |
6 tests, 0 passed, 6 failed
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Chasing The Leader :: Cluster Singleton test...
==============================================================================
Register_Candidates :: Register a candidate application on each no... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_chasing_the_leader_register_candidates"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_chasing_the_leader_register_candidates&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
Do_Nothing :: Do nothing for the time of the test duration, becaus... | PASS |
------------------------------------------------------------------------------
Unregister_Candidates_And_Validate_Criteria :: Unregister the test... | FAIL |
... click for list of related bugs or create a new one if needed (with the
"controller_clustering_tell_txt_chasing_the_leader_unregister_candidates_and_validate_criteria"
reference somewhere inside)
"https://bugs.opendaylight.org/buglist.cgi?f1=cf_external_ref&o1=substring&v1=controller_clustering_tell_txt_chasing_the_leader_unregister_candidates_and_validate_criteria&order=bug_status"
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Chasing The Leader :: Cluster Singl... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Restart Odl With Tell Based True :: Set tell...
==============================================================================
Stop_All_Members :: Stop every odl node. | FAIL |
Keyword 'Verify_Karaf_Is_Not_Running_On_Member' failed after retrying for 6 minutes. The last error was: Found running Karaf count: 1: 0 != 1
------------------------------------------------------------------------------
Set_Tell_Based_Protocol_Usage :: Un-comment the flag usage in conf... | PASS |
------------------------------------------------------------------------------
Start_All_And_Sync :: Start each member and wait for sync. | FAIL |
Keyword 'Verify_Members_Are_Ready' failed after retrying for 6 minutes. The last error was: ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=125.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Restart Odl With Tell Based True ::... | FAIL |
3 tests, 1 passed, 2 failed
==============================================================================
controller-clustering-tell.txt.Carpeople Crud :: Suite for performing basic...
==============================================================================
Add_Cars_To_Leader :: Add 30 cars to car Leader by one big PUT. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Added_Cars_On_Leader :: GET response from Leader should match ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Added_Cars_On_Followers :: The same check on other members. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Add_People_To_First_Follower :: Add 30 people to people first Foll... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Added_People_On_Leader :: GET response from Leader should matc... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Added_People_On_Followers :: The same check on other members. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Leader :: Buy some cars on car-people Leader, loop of ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Followers :: On each Follower buy corresponding ID seg... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Added_CarPeople_On_Leader :: GET car-person mappings from Lead... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Added_CarPeople_On_Followers :: The same check on other members. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_CarPeople_On_Leader :: DELETE car-people container. No ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_People_On_Leader :: DELETE people container. No verific... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_Cars_On_Leader :: DELETE cars container. No verificatio... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Carpeople Crud :: Suite for perform... | FAIL |
Suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
13 tests, 0 passed, 13 failed
==============================================================================
controller-clustering-tell.txt.Car Failover Crud :: Suite mixing basic oper...
==============================================================================
Add_Original_Cars_On_Old_Leader_And_Verify :: Add initial cars on ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Stop_Original_Car_Leader :: Stop the car Leader to cause a new lea... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Wait_For_New_Leader :: Wait until new car Leader is elected. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Original_Cars_On_New_Leader :: GET cars from new Leader, shoul... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Original_Cars_On_New_Followers :: The same check on other exis... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_Original_Cars_On_New_Leader :: Delete cars on the new Leader. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Add_Leader_Cars_On_New_Leader :: Add cars on the new Leader. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Leader_Cars_On_New_Leader :: GET cars from new Leader, should ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Leader_Cars_On_New_Followers :: The same check on other existi... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_Leader_Cars_On_New_First_Follower :: Delete cars in new fir... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Add_Follower_Cars_On_New_First_Follower :: Add cars on the new fir... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Folower_Cars_On_New_Leader :: Get cars from the new Leader, sh... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Follower_Cars_On_New_Followers :: The same check on other exis... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Start_Old_Car_Leader :: Start the stopped member without deleting ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Folower_Cars_On_Old_Leader :: GET cars from the restarted memb... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_Follower_Cars_On_New_Leader :: Delete cars on the last Leader. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Car Failover Crud :: Suite mixing b... | FAIL |
Suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
16 tests, 0 passed, 16 failed
==============================================================================
controller-clustering-tell.txt.Car Persistence Recovery :: This test restar...
==============================================================================
Add_Cars_On_Leader_And_Verify :: Single big PUT to datastore to ad... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Stop_All_Members :: Stop all controllers. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Start_All_Members :: Start all controllers (should restore the per... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Memorize_Leader_And_Followers :: Locate current Leader of car Shard. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Cars_On_Leader :: GET cars from Leader, should match the PUT d... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Cars_On_Followers :: The same check on other members. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_Cars_On_Leader :: Delete cars on the new Leader. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Car Persistence Recovery :: This te... | FAIL |
Suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
7 tests, 0 passed, 7 failed
==============================================================================
controller-clustering-tell.txt.Buycar Failover :: This test focuses on test...
==============================================================================
Add_Cars_To_Leader_And_Verify :: Add all needed cars to car Leader... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Add_People_To_First_Follower_And_Verify :: Add all needed people t... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Leader_And_Verify :: Buy some cars on the leader member. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Follower_And_Verify :: Buy some cars on the first foll... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Reboot_People_Leader :: Previous people Leader is rebooted. We sho... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Leader_After_Reboot_And_Verify :: Buy some cars on the... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Follower_After_Reboot_And_Verify :: Buy some cars on t... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_CarPeople :: DELETE car-people container. No verificati... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_People :: DELETE people container. No verification beyo... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_Cars :: DELETE cars container. No verification beyond h... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Buycar Failover :: This test focuse... | FAIL |
Suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
10 tests, 0 passed, 10 failed
==============================================================================
controller-clustering-tell.txt.Entity Isolate :: Suite for performing membe...
==============================================================================
Check All Shards Before Isolate :: Check all shards in controller. | FAIL |
Parent suite setup failed:
Evaluating expression 'json.loads(\'\'\'{\n "error": "javax.management.InstanceNotFoundException : org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "error_type": "javax.management.InstanceNotFoundException",\n "request": {\n "mbean": "org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "type": "read"\n },\n "stacktrace": "javax.management.InstanceNotFoundException: org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBean(DefaultMBeanServerInterceptor.java:1073)\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBeanInfo(DefaultMBeanServerInterceptor.java:1343)\\n\\tat java.management/com.sun.jmx.mbeanserver.JmxMBeanServer.getMBeanInfo(JmxMBeanServer.java:921)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:46)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:41)\\n\\tat org.jolokia.backend.executor.AbstractMBeanServerExecutor.call(AbstractMBeanServerExecutor.java:90)\\n\\tat org.jolokia.handler.ReadHandler.getMBeanInfo(ReadHandler.java:233)\\n\\tat org.jolokia.handler.ReadHandler.getAllAttributesNames(ReadHandler.java:245)\\n\\tat org.jolokia.handler.ReadHandler.resolveAttributes(ReadHandler.java:221)\\n\\tat org.jolokia.handler.ReadHandler.f...
[ Message content over the limit has been removed. ]
...lipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)\\n\\tat org.eclipse.jetty.server.Server.handle(Server.java:516)\\n\\tat org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:487)\\n\\tat org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:732)\\n\\tat org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:479)\\n\\tat org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277)\\n\\tat org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)\\n\\tat org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105)\\n\\tat org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)\\n\\tat org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)\\n\\tat java.base/java.lang.Thread.run(Thread.java:1583)\\n",\n "status": 404\n}\n\'\'\')' failed: JSONDecodeError: Invalid control character at: line 8 column 190 (char 619)
------------------------------------------------------------------------------
Isolate Entity Leader :: Isolate the entity-ownership Leader to ca... | FAIL |
Parent suite setup failed:
Evaluating expression 'json.loads(\'\'\'{\n "error": "javax.management.InstanceNotFoundException : org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "error_type": "javax.management.InstanceNotFoundException",\n "request": {\n "mbean": "org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "type": "read"\n },\n "stacktrace": "javax.management.InstanceNotFoundException: org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBean(DefaultMBeanServerInterceptor.java:1073)\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBeanInfo(DefaultMBeanServerInterceptor.java:1343)\\n\\tat java.management/com.sun.jmx.mbeanserver.JmxMBeanServer.getMBeanInfo(JmxMBeanServer.java:921)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:46)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:41)\\n\\tat org.jolokia.backend.executor.AbstractMBeanServerExecutor.call(AbstractMBeanServerExecutor.java:90)\\n\\tat org.jolokia.handler.ReadHandler.getMBeanInfo(ReadHandler.java:233)\\n\\tat org.jolokia.handler.ReadHandler.getAllAttributesNames(ReadHandler.java:245)\\n\\tat org.jolokia.handler.ReadHandler.resolveAttributes(ReadHandler.java:221)\\n\\tat org.jolokia.handler.ReadHandler.f...
[ Message content over the limit has been removed. ]
...lipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)\\n\\tat org.eclipse.jetty.server.Server.handle(Server.java:516)\\n\\tat org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:487)\\n\\tat org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:732)\\n\\tat org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:479)\\n\\tat org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277)\\n\\tat org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)\\n\\tat org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105)\\n\\tat org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)\\n\\tat org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)\\n\\tat java.base/java.lang.Thread.run(Thread.java:1583)\\n",\n "status": 404\n}\n\'\'\')' failed: JSONDecodeError: Invalid control character at: line 8 column 190 (char 619)
------------------------------------------------------------------------------
Check All Shards After Isolate :: Check all shards in controller. | FAIL |
Parent suite setup failed:
Evaluating expression 'json.loads(\'\'\'{\n "error": "javax.management.InstanceNotFoundException : org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "error_type": "javax.management.InstanceNotFoundException",\n "request": {\n "mbean": "org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "type": "read"\n },\n "stacktrace": "javax.management.InstanceNotFoundException: org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBean(DefaultMBeanServerInterceptor.java:1073)\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBeanInfo(DefaultMBeanServerInterceptor.java:1343)\\n\\tat java.management/com.sun.jmx.mbeanserver.JmxMBeanServer.getMBeanInfo(JmxMBeanServer.java:921)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:46)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:41)\\n\\tat org.jolokia.backend.executor.AbstractMBeanServerExecutor.call(AbstractMBeanServerExecutor.java:90)\\n\\tat org.jolokia.handler.ReadHandler.getMBeanInfo(ReadHandler.java:233)\\n\\tat org.jolokia.handler.ReadHandler.getAllAttributesNames(ReadHandler.java:245)\\n\\tat org.jolokia.handler.ReadHandler.resolveAttributes(ReadHandler.java:221)\\n\\tat org.jolokia.handler.ReadHandler.f...
[ Message content over the limit has been removed. ]
...lipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)\\n\\tat org.eclipse.jetty.server.Server.handle(Server.java:516)\\n\\tat org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:487)\\n\\tat org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:732)\\n\\tat org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:479)\\n\\tat org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277)\\n\\tat org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)\\n\\tat org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105)\\n\\tat org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)\\n\\tat org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)\\n\\tat java.base/java.lang.Thread.run(Thread.java:1583)\\n",\n "status": 404\n}\n\'\'\')' failed: JSONDecodeError: Invalid control character at: line 8 column 190 (char 619)
------------------------------------------------------------------------------
Rejoin Entity Leader :: Rejoin the entity-ownership Leader. | FAIL |
Parent suite setup failed:
Evaluating expression 'json.loads(\'\'\'{\n "error": "javax.management.InstanceNotFoundException : org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "error_type": "javax.management.InstanceNotFoundException",\n "request": {\n "mbean": "org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "type": "read"\n },\n "stacktrace": "javax.management.InstanceNotFoundException: org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBean(DefaultMBeanServerInterceptor.java:1073)\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBeanInfo(DefaultMBeanServerInterceptor.java:1343)\\n\\tat java.management/com.sun.jmx.mbeanserver.JmxMBeanServer.getMBeanInfo(JmxMBeanServer.java:921)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:46)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:41)\\n\\tat org.jolokia.backend.executor.AbstractMBeanServerExecutor.call(AbstractMBeanServerExecutor.java:90)\\n\\tat org.jolokia.handler.ReadHandler.getMBeanInfo(ReadHandler.java:233)\\n\\tat org.jolokia.handler.ReadHandler.getAllAttributesNames(ReadHandler.java:245)\\n\\tat org.jolokia.handler.ReadHandler.resolveAttributes(ReadHandler.java:221)\\n\\tat org.jolokia.handler.ReadHandler.f...
[ Message content over the limit has been removed. ]
...lipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)\\n\\tat org.eclipse.jetty.server.Server.handle(Server.java:516)\\n\\tat org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:487)\\n\\tat org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:732)\\n\\tat org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:479)\\n\\tat org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277)\\n\\tat org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)\\n\\tat org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105)\\n\\tat org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)\\n\\tat org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)\\n\\tat java.base/java.lang.Thread.run(Thread.java:1583)\\n",\n "status": 404\n}\n\'\'\')' failed: JSONDecodeError: Invalid control character at: line 8 column 190 (char 619)
------------------------------------------------------------------------------
Check All Shards After Rejoin :: Check all shards in controller. | FAIL |
Parent suite setup failed:
Evaluating expression 'json.loads(\'\'\'{\n "error": "javax.management.InstanceNotFoundException : org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "error_type": "javax.management.InstanceNotFoundException",\n "request": {\n "mbean": "org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "type": "read"\n },\n "stacktrace": "javax.management.InstanceNotFoundException: org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBean(DefaultMBeanServerInterceptor.java:1073)\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBeanInfo(DefaultMBeanServerInterceptor.java:1343)\\n\\tat java.management/com.sun.jmx.mbeanserver.JmxMBeanServer.getMBeanInfo(JmxMBeanServer.java:921)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:46)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:41)\\n\\tat org.jolokia.backend.executor.AbstractMBeanServerExecutor.call(AbstractMBeanServerExecutor.java:90)\\n\\tat org.jolokia.handler.ReadHandler.getMBeanInfo(ReadHandler.java:233)\\n\\tat org.jolokia.handler.ReadHandler.getAllAttributesNames(ReadHandler.java:245)\\n\\tat org.jolokia.handler.ReadHandler.resolveAttributes(ReadHandler.java:221)\\n\\tat org.jolokia.handler.ReadHandler.f...
[ Message content over the limit has been removed. ]
...lipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)\\n\\tat org.eclipse.jetty.server.Server.handle(Server.java:516)\\n\\tat org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:487)\\n\\tat org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:732)\\n\\tat org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:479)\\n\\tat org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277)\\n\\tat org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)\\n\\tat org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105)\\n\\tat org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)\\n\\tat org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)\\n\\tat java.base/java.lang.Thread.run(Thread.java:1583)\\n",\n "status": 404\n}\n\'\'\')' failed: JSONDecodeError: Invalid control character at: line 8 column 190 (char 619)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Entity Isolate :: Suite for perform... | FAIL |
Suite setup failed:
Evaluating expression 'json.loads(\'\'\'{\n "error": "javax.management.InstanceNotFoundException : org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "error_type": "javax.management.InstanceNotFoundException",\n "request": {\n "mbean": "org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore",\n "type": "read"\n },\n "stacktrace": "javax.management.InstanceNotFoundException: org.opendaylight.controller:Category=Shards,name=member-1-shard-entity-ownership-operational,type=DistributedOperationalDatastore\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBean(DefaultMBeanServerInterceptor.java:1073)\\n\\tat java.management/com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getMBeanInfo(DefaultMBeanServerInterceptor.java:1343)\\n\\tat java.management/com.sun.jmx.mbeanserver.JmxMBeanServer.getMBeanInfo(JmxMBeanServer.java:921)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:46)\\n\\tat org.jolokia.handler.ReadHandler$1.execute(ReadHandler.java:41)\\n\\tat org.jolokia.backend.executor.AbstractMBeanServerExecutor.call(AbstractMBeanServerExecutor.java:90)\\n\\tat org.jolokia.handler.ReadHandler.getMBeanInfo(ReadHandler.java:233)\\n\\tat org.jolokia.handler.ReadHandler.getAllAttributesNames(ReadHandler.java:245)\\n\\tat org.jolokia.handler.ReadHandler.resolveAttributes(ReadHandler.java:221)\\n\\tat org.jolokia.handler.ReadHandler.f...
[ Message content over the limit has been removed. ]
...lipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)\\n\\tat org.eclipse.jetty.server.Server.handle(Server.java:516)\\n\\tat org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:487)\\n\\tat org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:732)\\n\\tat org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:479)\\n\\tat org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277)\\n\\tat org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)\\n\\tat org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105)\\n\\tat org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)\\n\\tat org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)\\n\\tat org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)\\n\\tat org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)\\n\\tat java.base/java.lang.Thread.run(Thread.java:1583)\\n",\n "status": 404\n}\n\'\'\')' failed: JSONDecodeError: Invalid control character at: line 8 column 190 (char 619)
5 tests, 0 passed, 5 failed
==============================================================================
controller-clustering-tell.txt.Buycar Failover Isolation :: This test focus...
==============================================================================
Add_Cars_To_Leader_And_Verify :: Add all needed cars to car Leader... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Add_People_To_First_Follower_And_Verify :: Add all needed people t... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Leader_And_Verify :: Buy some cars on the leader member. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Follower_And_Verify :: Buy some cars on the first foll... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Isolate_and_Rejoin_People_Leader :: Previous people Leader is isol... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Leader_After_Rejoin_And_Verify :: Buy some cars on the... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Buy_Cars_On_Follower_After_Rejoin_And_Verify :: Buy some cars on t... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_CarPeople :: DELETE car-people container. No verificati... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_People :: DELETE people container. No verification beyo... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_All_Cars :: DELETE cars container. No verification beyond h... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Buycar Failover Isolation :: This t... | FAIL |
Suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
10 tests, 0 passed, 10 failed
==============================================================================
controller-clustering-tell.txt.Car Failover Crud Isolation :: Suite mixing ...
==============================================================================
Add_Original_Cars_On_Old_Leader_And_Verify :: Add initial cars on ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Isolate_Original_Car_Leader :: Isolate the car Leader to cause a n... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Wait_For_New_Leader :: Wait until new car Leader is elected. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Original_Cars_On_New_Leader :: GET cars from new Leader, shoul... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Original_Cars_On_New_Followers :: The same check on other exis... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_Original_Cars_On_New_Leader :: Delete cars on the new Leader. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Add_Leader_Cars_On_New_Leader :: Add cars on the new Leader. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Leader_Cars_On_New_Leader :: GET cars from new Leader, should ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Leader_Cars_On_New_Followers :: The same check on other existi... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_Leader_Cars_On_New_First_Follower :: Delete cars in new fir... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Add_Follower_Cars_On_New_First_Follower :: Add cars on the new fir... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Folower_Cars_On_New_Leader :: Get cars from the new Leader, sh... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Follower_Cars_On_New_Followers :: The same check on other exis... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Rejoin_Old_Car_Leader :: Rejoin the isolated member without deleti... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Folower_Cars_On_Old_Leader :: GET cars from the restarted memb... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_Follower_Cars_On_New_Leader :: Delete cars on the last Leader. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Car Failover Crud Isolation :: Suit... | FAIL |
Suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
16 tests, 0 passed, 16 failed
==============================================================================
controller-clustering-tell.txt.Car Outage Corners :: Cluster suite for test...
==============================================================================
Stop_Majority_Of_The_Followers :: Stop half plus one car Follower ... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Attempt_To_Add_Cars_To_Leader :: Adding cars should fail, as major... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Start_Tipping_Follower :: Start one Follower member without persis... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Add_Cars_On_Tipping_Follower :: Add cars on the tipping Follower. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Cars_On_Existing_Members :: On each up member: GET cars, shoul... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Start_Other_Followers :: Start other followers without persisted d... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
See_Cars_On_New_Follower_Leader :: GET cars from a new follower to... | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
Delete_Cars_On_Leader :: Delete cars on Leader. | FAIL |
Parent suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
------------------------------------------------------------------------------
controller-clustering-tell.txt.Car Outage Corners :: Cluster suite... | FAIL |
Suite setup failed:
ReadTimeout: HTTPConnectionPool(host='10.30.171.180', port=8181): Read timed out. (read timeout=5.0)
8 tests, 0 passed, 8 failed
==============================================================================
controller-clustering-tell.txt | FAIL |
162 tests, 11 passed, 151 failed
==============================================================================
Output: /w/workspace/controller-csit-3node-clustering-tell-all-scandium/output.xml
Log: /w/workspace/controller-csit-3node-clustering-tell-all-scandium/log.html
Report: /w/workspace/controller-csit-3node-clustering-tell-all-scandium/report.html
+ true
+ echo 'Examining the files in data/log and checking filesize'
Examining the files in data/log and checking filesize
+ ssh 10.30.170.103 'ls -altr /tmp/karaf-0.21.3-SNAPSHOT/data/log/'
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
total 8428
drwxrwxr-x 2 jenkins jenkins 4096 Oct 6 19:31 .
-rw-rw-r-- 1 jenkins jenkins 1720 Oct 6 19:31 karaf_console.log
drwxrwxr-x 9 jenkins jenkins 4096 Oct 7 00:19 ..
-rw-rw-r-- 1 jenkins jenkins 8612026 Oct 7 00:27 karaf.log
+ ssh 10.30.170.103 'du -hs /tmp/karaf-0.21.3-SNAPSHOT/data/log/*'
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
8.3M /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log
4.0K /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
+ ssh 10.30.171.180 'ls -altr /tmp/karaf-0.21.3-SNAPSHOT/data/log/'
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
total 6296
drwxrwxr-x 2 jenkins jenkins 4096 Oct 6 19:31 .
-rw-rw-r-- 1 jenkins jenkins 8319 Oct 6 19:41 karaf_console.log
drwxrwxr-x 4 jenkins jenkins 4096 Oct 7 00:19 ..
-rw-rw-r-- 1 jenkins jenkins 6419386 Oct 7 00:27 karaf.log
+ ssh 10.30.171.180 'du -hs /tmp/karaf-0.21.3-SNAPSHOT/data/log/*'
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
6.2M /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log
12K /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
+ ssh 10.30.171.153 'ls -altr /tmp/karaf-0.21.3-SNAPSHOT/data/log/'
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
total 2824
drwxrwxr-x 2 jenkins jenkins 4096 Oct 6 19:31 .
-rw-rw-r-- 1 jenkins jenkins 132599 Oct 6 20:02 karaf_console.log
drwxrwxr-x 4 jenkins jenkins 4096 Oct 7 00:19 ..
-rw-rw-r-- 1 jenkins jenkins 2747635 Oct 7 00:27 karaf.log
+ ssh 10.30.171.153 'du -hs /tmp/karaf-0.21.3-SNAPSHOT/data/log/*'
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
2.7M /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log
132K /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
+ set +e
++ seq 1 3
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_1_IP
+ echo 'Let'\''s take the karaf thread dump again'
Let's take the karaf thread dump again
+ ssh 10.30.170.103 'sudo ps aux'
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
++ grep org.apache.karaf.main.Main /w/workspace/controller-csit-3node-clustering-tell-all-scandium/ps_after.log
++ grep -v grep
++ tr -s ' '
++ cut -f2 '-d '
+ pid=29650
+ echo 'karaf main: org.apache.karaf.main.Main, pid:29650'
karaf main: org.apache.karaf.main.Main, pid:29650
+ ssh 10.30.170.103 '/usr/lib/jvm/java-21-openjdk-amd64/bin/jstack -l 29650'
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ echo 'killing karaf process...'
killing karaf process...
+ ssh 10.30.170.103 bash -c 'ps axf | grep karaf | grep -v grep | awk '\''{print "kill -9 " $1}'\'' | sh'
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_2_IP
+ echo 'Let'\''s take the karaf thread dump again'
Let's take the karaf thread dump again
+ ssh 10.30.171.180 'sudo ps aux'
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
++ grep org.apache.karaf.main.Main /w/workspace/controller-csit-3node-clustering-tell-all-scandium/ps_after.log
++ grep -v grep
++ tr -s ' '
++ cut -f2 '-d '
+ pid=2080
+ echo 'karaf main: org.apache.karaf.main.Main, pid:2080'
karaf main: org.apache.karaf.main.Main, pid:2080
+ ssh 10.30.171.180 '/usr/lib/jvm/java-21-openjdk-amd64/bin/jstack -l 2080'
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ echo 'killing karaf process...'
killing karaf process...
+ ssh 10.30.171.180 bash -c 'ps axf | grep karaf | grep -v grep | awk '\''{print "kill -9 " $1}'\'' | sh'
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_3_IP
+ echo 'Let'\''s take the karaf thread dump again'
Let's take the karaf thread dump again
+ ssh 10.30.171.153 'sudo ps aux'
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
++ grep org.apache.karaf.main.Main /w/workspace/controller-csit-3node-clustering-tell-all-scandium/ps_after.log
++ grep -v grep
++ tr -s ' '
++ cut -f2 '-d '
+ pid=2082
+ echo 'karaf main: org.apache.karaf.main.Main, pid:2082'
karaf main: org.apache.karaf.main.Main, pid:2082
+ ssh 10.30.171.153 '/usr/lib/jvm/java-21-openjdk-amd64/bin/jstack -l 2082'
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ echo 'killing karaf process...'
killing karaf process...
+ ssh 10.30.171.153 bash -c 'ps axf | grep karaf | grep -v grep | awk '\''{print "kill -9 " $1}'\'' | sh'
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ sleep 5
++ seq 1 3
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_1_IP
+ echo 'Compressing karaf.log 1'
Compressing karaf.log 1
+ ssh 10.30.170.103 gzip --best /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ echo 'Fetching compressed karaf.log 1'
Fetching compressed karaf.log 1
+ scp 10.30.170.103:/tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log.gz odl1_karaf.log.gz
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ ssh 10.30.170.103 rm -f /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log.gz
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ scp 10.30.170.103:/tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log odl1_karaf_console.log
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ ssh 10.30.170.103 rm -f /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
+ echo 'Fetch GC logs'
Fetch GC logs
+ mkdir -p gclogs-1
+ scp '10.30.170.103:/tmp/karaf-0.21.3-SNAPSHOT/data/log/*.log' gclogs-1/
Warning: Permanently added '10.30.170.103' (ECDSA) to the list of known hosts.
scp: /tmp/karaf-0.21.3-SNAPSHOT/data/log/*.log: No such file or directory
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_2_IP
+ echo 'Compressing karaf.log 2'
Compressing karaf.log 2
+ ssh 10.30.171.180 gzip --best /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ echo 'Fetching compressed karaf.log 2'
Fetching compressed karaf.log 2
+ scp 10.30.171.180:/tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log.gz odl2_karaf.log.gz
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ ssh 10.30.171.180 rm -f /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log.gz
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ scp 10.30.171.180:/tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log odl2_karaf_console.log
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ ssh 10.30.171.180 rm -f /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
+ echo 'Fetch GC logs'
Fetch GC logs
+ mkdir -p gclogs-2
+ scp '10.30.171.180:/tmp/karaf-0.21.3-SNAPSHOT/data/log/*.log' gclogs-2/
Warning: Permanently added '10.30.171.180' (ECDSA) to the list of known hosts.
scp: /tmp/karaf-0.21.3-SNAPSHOT/data/log/*.log: No such file or directory
+ for i in $(seq 1 "${NUM_ODL_SYSTEM}")
+ CONTROLLERIP=ODL_SYSTEM_3_IP
+ echo 'Compressing karaf.log 3'
Compressing karaf.log 3
+ ssh 10.30.171.153 gzip --best /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ echo 'Fetching compressed karaf.log 3'
Fetching compressed karaf.log 3
+ scp 10.30.171.153:/tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log.gz odl3_karaf.log.gz
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ ssh 10.30.171.153 rm -f /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf.log.gz
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ scp 10.30.171.153:/tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log odl3_karaf_console.log
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ ssh 10.30.171.153 rm -f /tmp/karaf-0.21.3-SNAPSHOT/data/log/karaf_console.log
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
+ echo 'Fetch GC logs'
Fetch GC logs
+ mkdir -p gclogs-3
+ scp '10.30.171.153:/tmp/karaf-0.21.3-SNAPSHOT/data/log/*.log' gclogs-3/
Warning: Permanently added '10.30.171.153' (ECDSA) to the list of known hosts.
scp: /tmp/karaf-0.21.3-SNAPSHOT/data/log/*.log: No such file or directory
+ echo 'Examine copied files'
Examine copied files
+ ls -lt
total 68168
drwxrwxr-x. 2 jenkins jenkins 6 Oct 7 00:27 gclogs-3
-rw-rw-r--. 1 jenkins jenkins 132599 Oct 7 00:27 odl3_karaf_console.log
-rw-rw-r--. 1 jenkins jenkins 133350 Oct 7 00:27 odl3_karaf.log.gz
drwxrwxr-x. 2 jenkins jenkins 6 Oct 7 00:27 gclogs-2
-rw-rw-r--. 1 jenkins jenkins 8319 Oct 7 00:27 odl2_karaf_console.log
-rw-rw-r--. 1 jenkins jenkins 225764 Oct 7 00:27 odl2_karaf.log.gz
drwxrwxr-x. 2 jenkins jenkins 6 Oct 7 00:27 gclogs-1
-rw-rw-r--. 1 jenkins jenkins 1720 Oct 7 00:27 odl1_karaf_console.log
-rw-rw-r--. 1 jenkins jenkins 622443 Oct 7 00:27 odl1_karaf.log.gz
-rw-rw-r--. 1 jenkins jenkins 131890 Oct 7 00:27 karaf_3_2082_threads_after.log
-rw-rw-r--. 1 jenkins jenkins 13549 Oct 7 00:27 ps_after.log
-rw-rw-r--. 1 jenkins jenkins 131080 Oct 7 00:27 karaf_2_2080_threads_after.log
-rw-rw-r--. 1 jenkins jenkins 134256 Oct 7 00:27 karaf_1_29650_threads_after.log
-rw-rw-r--. 1 jenkins jenkins 307579 Oct 7 00:27 report.html
-rw-rw-r--. 1 jenkins jenkins 3433450 Oct 7 00:27 log.html
-rw-rw-r--. 1 jenkins jenkins 64067751 Oct 7 00:27 output.xml
-rw-rw-r--. 1 jenkins jenkins 6069 Oct 6 19:34 testplan.txt
-rw-rw-r--. 1 jenkins jenkins 110588 Oct 6 19:34 karaf_3_2082_threads_before.log
-rw-rw-r--. 1 jenkins jenkins 14299 Oct 6 19:34 ps_before.log
-rw-rw-r--. 1 jenkins jenkins 110611 Oct 6 19:34 karaf_2_2080_threads_before.log
-rw-rw-r--. 1 jenkins jenkins 116901 Oct 6 19:34 karaf_1_2085_threads_before.log
-rw-rw-r--. 1 jenkins jenkins 3106 Oct 6 19:31 post-startup-script.sh
-rw-rw-r--. 1 jenkins jenkins 252 Oct 6 19:31 startup-script.sh
-rw-rw-r--. 1 jenkins jenkins 3472 Oct 6 19:31 configuration-script.sh
-rw-rw-r--. 1 jenkins jenkins 695 Oct 6 19:31 custom_shard_config.txt
-rw-rw-r--. 1 jenkins jenkins 136 Oct 6 19:31 scriptplan.txt
-rw-rw-r--. 1 jenkins jenkins 337 Oct 6 19:30 detect_variables.env
-rw-rw-r--. 1 jenkins jenkins 2619 Oct 6 19:30 pom.xml
-rw-rw-r--. 1 jenkins jenkins 92 Oct 6 19:30 set_variables.env
-rw-rw-r--. 1 jenkins jenkins 314 Oct 6 19:30 slave_addresses.txt
-rw-rw-r--. 1 jenkins jenkins 570 Oct 6 19:30 requirements.txt
-rw-rw-r--. 1 jenkins jenkins 26 Oct 6 19:30 env.properties
-rw-rw-r--. 1 jenkins jenkins 333 Oct 6 19:28 stack-parameters.yaml
drwxrwxr-x. 7 jenkins jenkins 4096 Oct 6 19:27 test
drwxrwxr-x. 2 jenkins jenkins 6 Oct 6 19:27 test@tmp
-rw-rw-r--. 1 jenkins jenkins 1410 Oct 6 18:35 maven-metadata.xml
+ true
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/sh /tmp/jenkins15430762198553025864.sh
Cleaning up Robot installation...
$ ssh-agent -k
unset SSH_AUTH_SOCK;
unset SSH_AGENT_PID;
echo Agent pid 5163 killed;
[ssh-agent] Stopped.
Recording plot data
Robot results publisher started...
INFO: Checking test criticality is deprecated and will be dropped in a future release!
-Parsing output xml:
Done!
-Copying log files to build dir:
Done!
-Assigning results to build:
Done!
-Checking thresholds:
Done!
Done publishing Robot results.
Build step 'Publish Robot Framework test results' changed build result to UNSTABLE
[PostBuildScript] - [INFO] Executing post build scripts.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins6939763109241909340.sh
Archiving csit artifacts
mv: cannot stat '*_1.png': No such file or directory
mv: cannot stat '/tmp/odl1_*': No such file or directory
mv: cannot stat '*_2.png': No such file or directory
mv: cannot stat '/tmp/odl2_*': No such file or directory
mv: cannot stat '*_3.png': No such file or directory
mv: cannot stat '/tmp/odl3_*': No such file or directory
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
100 1981k 0 1981k 0 0 4011k 0 --:--:-- --:--:-- --:--:-- 4003k
100 4998k 0 4998k 0 0 4454k 0 --:--:-- 0:00:01 --:--:-- 4454k
Archive: robot-plugin.zip
inflating: ./archives/robot-plugin/log.html
inflating: ./archives/robot-plugin/output.xml
inflating: ./archives/robot-plugin/report.html
mv: cannot stat '*.log.gz': No such file or directory
mv: cannot stat '*.csv': No such file or directory
mv: cannot stat '*.png': No such file or directory
[PostBuildScript] - [INFO] Executing post build scripts.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins16340756080327326244.sh
[PostBuildScript] - [INFO] Executing post build scripts.
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content
OS_CLOUD=vex
OS_STACK_NAME=releng-controller-csit-3node-clustering-tell-all-scandium-76
[EnvInject] - Variables injected successfully.
provisioning config files...
copy managed file [clouds-yaml] to file:/home/jenkins/.config/openstack/clouds.yaml
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins395458198050934928.sh
---> openstack-stack-delete.sh
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
lftools 0.37.14 requires urllib3<2.1.0, but you have urllib3 2.5.0 which is incompatible.
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: lftools[openstack] kubernetes python-heatclient python-openstackclient urllib3~=1.26.15
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
INFO: Retrieving stack cost for: releng-controller-csit-3node-clustering-tell-all-scandium-76
DEBUG: Successfully retrieved stack cost: total: 1.6500000000000001
INFO: Deleting stack releng-controller-csit-3node-clustering-tell-all-scandium-76
Successfully deleted stack releng-controller-csit-3node-clustering-tell-all-scandium-76
[PostBuildScript] - [INFO] Executing post build scripts.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins15016405414580215605.sh
---> sysstat.sh
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins14142563563832948041.sh
---> package-listing.sh
++ facter osfamily
++ tr '[:upper:]' '[:lower:]'
+ OS_FAMILY=redhat
+ workspace=/w/workspace/controller-csit-3node-clustering-tell-all-scandium
+ START_PACKAGES=/tmp/packages_start.txt
+ END_PACKAGES=/tmp/packages_end.txt
+ DIFF_PACKAGES=/tmp/packages_diff.txt
+ PACKAGES=/tmp/packages_start.txt
+ '[' /w/workspace/controller-csit-3node-clustering-tell-all-scandium ']'
+ PACKAGES=/tmp/packages_end.txt
+ case "${OS_FAMILY}" in
+ rpm -qa
+ sort
+ '[' -f /tmp/packages_start.txt ']'
+ '[' -f /tmp/packages_end.txt ']'
+ diff /tmp/packages_start.txt /tmp/packages_end.txt
+ '[' /w/workspace/controller-csit-3node-clustering-tell-all-scandium ']'
+ mkdir -p /w/workspace/controller-csit-3node-clustering-tell-all-scandium/archives/
+ cp -f /tmp/packages_diff.txt /tmp/packages_end.txt /tmp/packages_start.txt /w/workspace/controller-csit-3node-clustering-tell-all-scandium/archives/
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins3435363206820361630.sh
---> capture-instance-metadata.sh
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: lftools
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
INFO: Running in OpenStack, capturing instance metadata
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins10395059944188460566.sh
provisioning config files...
Could not find credentials [logs] for controller-csit-3node-clustering-tell-all-scandium #76
copy managed file [jenkins-log-archives-settings] to file:/w/workspace/controller-csit-3node-clustering-tell-all-scandium@tmp/config16478849052298166375tmp
Regular expression run condition: Expression=[^.*logs-s3.*], Label=[odl-logs-s3-cloudfront-index]
Run condition [Regular expression match] enabling perform for step [Provide Configuration files]
provisioning config files...
copy managed file [jenkins-s3-log-ship] to file:/home/jenkins/.aws/credentials
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content
SERVER_ID=logs
[EnvInject] - Variables injected successfully.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins11361992379865508391.sh
---> create-netrc.sh
WARN: Log server credential not found.
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins5620648928957875457.sh
---> python-tools-install.sh
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: lftools
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins17640151920294730404.sh
---> sudo-logs.sh
Archiving 'sudo' log..
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash /tmp/jenkins13357470942081841831.sh
---> job-cost.sh
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: zipp==1.1.0 python-openstackclient urllib3~=1.26.15
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
DEBUG: total: 1.6500000000000001
INFO: Retrieving Stack Cost...
INFO: Retrieving Pricing Info for: v3-standard-2
INFO: Archiving Costs
[controller-csit-3node-clustering-tell-all-scandium] $ /bin/bash -l /tmp/jenkins1626914728960327807.sh
---> logs-deploy.sh
Setup pyenv:
system
3.8.13
3.9.13
3.10.13
* 3.11.7 (set by /w/workspace/controller-csit-3node-clustering-tell-all-scandium/.python-version)
lf-activate-venv(): INFO: Reuse venv:/tmp/venv-DyLm from file:/tmp/.os_lf_venv
lf-activate-venv(): INFO: Installing base packages (pip, setuptools, virtualenv)
lf-activate-venv(): INFO: Attempting to install with network-safe options...
lf-activate-venv(): INFO: Base packages installed successfully
lf-activate-venv(): INFO: Installing additional packages: lftools urllib3~=1.26.15
lf-activate-venv(): INFO: Adding /tmp/venv-DyLm/bin to PATH
WARNING: Nexus logging server not set
INFO: S3 path logs/releng/vex-yul-odl-jenkins-1/controller-csit-3node-clustering-tell-all-scandium/76/
INFO: archiving logs to S3
---> uname -a:
Linux prd-centos8-robot-2c-8g-1207.novalocal 4.18.0-553.5.1.el8.x86_64 #1 SMP Tue May 21 05:46:01 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
---> lscpu:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Byte Order: Little Endian
CPU(s): 2
On-line CPU(s) list: 0,1
Thread(s) per core: 1
Core(s) per socket: 1
Socket(s): 2
NUMA node(s): 1
Vendor ID: AuthenticAMD
CPU family: 23
Model: 49
Model name: AMD EPYC-Rome Processor
Stepping: 0
CPU MHz: 2799.998
BogoMIPS: 5599.99
Virtualization: AMD-V
Hypervisor vendor: KVM
Virtualization type: full
L1d cache: 32K
L1i cache: 32K
L2 cache: 512K
L3 cache: 16384K
NUMA node0 CPU(s): 0,1
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm rep_good nopl cpuid extd_apicid tsc_known_freq pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy svm cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext perfctr_core ssbd ibrs ibpb stibp vmmcall fsgsbase tsc_adjust bmi1 avx2 smep bmi2 rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr wbnoinvd arat npt nrip_save umip rdpid arch_capabilities
---> nproc:
2
---> df -h:
Filesystem Size Used Avail Use% Mounted on
devtmpfs 3.8G 0 3.8G 0% /dev
tmpfs 3.8G 0 3.8G 0% /dev/shm
tmpfs 3.8G 17M 3.8G 1% /run
tmpfs 3.8G 0 3.8G 0% /sys/fs/cgroup
/dev/vda1 40G 8.5G 32G 22% /
tmpfs 770M 0 770M 0% /run/user/1001
---> free -m:
total used free shared buff/cache available
Mem: 7697 636 4716 19 2345 6763
Swap: 1023 0 1023
---> ip addr:
1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: eth0: mtu 1458 qdisc mq state UP group default qlen 1000
link/ether fa:16:3e:c4:6c:5d brd ff:ff:ff:ff:ff:ff
altname enp0s3
altname ens3
inet 10.30.170.85/23 brd 10.30.171.255 scope global dynamic noprefixroute eth0
valid_lft 68257sec preferred_lft 68257sec
inet6 fe80::f816:3eff:fec4:6c5d/64 scope link
valid_lft forever preferred_lft forever
---> sar -b -r -n DEV:
Linux 4.18.0-553.5.1.el8.x86_64 (prd-centos8-robot-2c-8g-1207.novalocal) 10/07/2025 _x86_64_ (2 CPU)
12:00:01 AM tps rtps wtps bread/s bwrtn/s
12:01:01 AM 0.98 0.00 0.98 0.00 216.61
12:02:01 AM 0.58 0.00 0.58 0.00 7.88
12:03:01 AM 0.43 0.00 0.43 0.00 6.10
12:04:01 AM 0.37 0.00 0.37 0.00 5.33
12:05:01 AM 0.30 0.00 0.30 0.00 3.22
12:06:01 AM 0.30 0.00 0.30 0.00 5.23
12:07:01 AM 0.22 0.00 0.22 0.00 2.37
12:08:01 AM 0.52 0.07 0.45 7.60 36.07
12:09:01 AM 0.32 0.00 0.32 0.00 4.85
12:10:01 AM 0.12 0.00 0.12 0.00 0.80
12:11:01 AM 1.10 0.00 1.10 0.00 19.66
12:12:01 AM 0.32 0.00 0.32 0.00 8.78
12:13:01 AM 0.20 0.00 0.20 0.00 3.32
12:14:01 AM 0.23 0.00 0.23 0.00 23.08
12:15:01 AM 0.52 0.00 0.52 0.00 17.86
12:16:01 AM 0.52 0.00 0.52 0.00 19.08
12:17:01 AM 0.28 0.00 0.28 0.00 15.56
12:18:01 AM 0.47 0.00 0.47 0.00 17.61
12:19:01 AM 0.28 0.00 0.28 0.00 17.74
12:20:01 AM 0.18 0.00 0.18 0.00 19.06
12:21:01 AM 0.43 0.00 0.43 0.00 5.13
12:22:01 AM 0.37 0.00 0.37 0.00 5.05
12:23:01 AM 0.57 0.00 0.57 0.00 8.21
12:24:01 AM 0.38 0.00 0.38 0.00 4.27
12:25:01 AM 0.33 0.00 0.33 0.00 5.30
12:26:01 AM 0.32 0.00 0.32 0.00 3.28
12:27:01 AM 0.38 0.00 0.38 0.00 45.42
12:28:01 AM 15.68 0.28 15.40 33.86 3610.53
12:29:01 AM 46.42 0.43 45.98 49.18 2331.87
Average: 2.52 0.03 2.50 3.13 223.20
12:00:01 AM kbmemfree kbavail kbmemused %memused kbbuffers kbcached kbcommit %commit kbactive kbinact kbdirty
12:01:01 AM 4801880 6981204 3080552 39.08 2688 2346632 759288 8.50 281664 2416236 8
12:02:01 AM 4803224 6982556 3079208 39.06 2688 2346640 759288 8.50 281680 2415280 4
12:03:01 AM 4803248 6982588 3079184 39.06 2688 2346648 759288 8.50 281680 2415216 4
12:04:01 AM 4803356 6982736 3079076 39.06 2688 2346688 759288 8.50 281680 2415192 4
12:05:01 AM 4803600 6982984 3078832 39.06 2688 2346692 759288 8.50 281680 2414980 4
12:06:01 AM 4803300 6982724 3079132 39.06 2688 2346732 759288 8.50 281680 2415260 4
12:07:01 AM 4803220 6982644 3079212 39.06 2688 2346732 758648 8.49 281680 2415288 4
12:08:01 AM 4802084 6982604 3080348 39.08 2688 2347828 744040 8.33 282096 2415916 4
12:09:01 AM 4802320 6982840 3080112 39.08 2688 2347828 744040 8.33 282096 2415848 4
12:10:01 AM 4802012 6982600 3080420 39.08 2688 2347896 742260 8.31 282820 2415344 324
12:11:01 AM 4802052 6982768 3080380 39.08 2688 2348024 742260 8.31 282824 2415364 128
12:12:01 AM 4802016 6982732 3080416 39.08 2688 2348024 742260 8.31 282824 2415364 4
12:13:01 AM 4801476 6982472 3080956 39.09 2688 2348312 760300 8.51 282824 2415892 296
12:14:01 AM 4801340 6982736 3081092 39.09 2688 2348712 760300 8.51 282824 2415944 52
12:15:01 AM 4801176 6982948 3081256 39.09 2688 2349080 760300 8.51 282824 2416432 24
12:16:01 AM 4800864 6983028 3081568 39.09 2688 2349480 760300 8.51 282824 2416708 24
12:17:01 AM 4800396 6982940 3082036 39.10 2688 2349860 760300 8.51 282824 2417088 24
12:18:01 AM 4799872 6982796 3082560 39.11 2688 2350240 760300 8.51 282824 2417572 96
12:19:01 AM 4799448 6982832 3082984 39.11 2688 2350692 760300 8.51 282824 2417964 104
12:20:01 AM 4798836 6982656 3083596 39.12 2688 2351128 750456 8.40 282824 2418632 4
12:21:01 AM 4798788 6982612 3083644 39.12 2688 2351132 750456 8.40 282824 2418516 4
12:22:01 AM 4798820 6982692 3083612 39.12 2688 2351180 750456 8.40 282824 2418560 4
12:23:01 AM 4798948 6982820 3083484 39.12 2688 2351180 750456 8.40 282824 2418632 4
12:24:01 AM 4798848 6982724 3083584 39.12 2688 2351192 766168 8.58 282824 2418456 4
12:25:01 AM 4798728 6982644 3083704 39.12 2688 2351224 766168 8.58 282824 2418820 4
12:26:01 AM 4799000 6982916 3083432 39.12 2688 2351224 766168 8.58 282824 2418540 4
12:27:01 AM 4790292 6976196 3092140 39.23 2688 2353212 883160 9.89 282828 2426208 788
12:28:01 AM 4752960 6893972 3129472 39.70 2688 2314324 771632 8.64 375544 2394056 59536
12:29:01 AM 4836460 6935744 3045972 38.64 2688 2272512 717496 8.03 543032 2145596 18648
Average: 4800295 6977783 3082137 39.10 2688 2345209 759447 8.50 294667 2406859 2763
12:00:01 AM IFACE rxpck/s txpck/s rxkB/s txkB/s rxcmp/s txcmp/s rxmcst/s %ifutil
12:01:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:01:01 AM eth0 15.09 12.21 3.98 2.46 0.00 0.00 0.00 0.00
12:02:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:02:01 AM eth0 0.77 0.17 0.28 0.19 0.00 0.00 0.00 0.00
12:03:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:03:01 AM eth0 0.95 0.23 0.18 0.07 0.00 0.00 0.00 0.00
12:04:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:04:01 AM eth0 0.62 0.13 0.06 0.02 0.00 0.00 0.00 0.00
12:05:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:05:01 AM eth0 0.47 0.00 0.03 0.00 0.00 0.00 0.00 0.00
12:06:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:06:01 AM eth0 1.62 0.47 0.41 0.26 0.00 0.00 0.00 0.00
12:07:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:07:01 AM eth0 0.47 0.00 0.06 0.00 0.00 0.00 0.00 0.00
12:08:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:08:01 AM eth0 17.58 16.56 3.15 1.60 0.00 0.00 0.00 0.00
12:09:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:09:01 AM eth0 0.52 0.07 0.03 0.00 0.00 0.00 0.00 0.00
12:10:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:10:01 AM eth0 6.39 5.88 0.87 0.42 0.00 0.00 0.00 0.00
12:11:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:11:01 AM eth0 6.68 6.13 1.01 0.43 0.00 0.00 0.00 0.00
12:12:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:12:01 AM eth0 0.45 0.03 0.03 0.00 0.00 0.00 0.00 0.00
12:13:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:13:01 AM eth0 16.26 15.05 3.19 1.83 0.00 0.00 0.00 0.00
12:14:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:14:01 AM eth0 11.05 9.10 2.79 1.79 0.00 0.00 0.00 0.00
12:15:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:15:01 AM eth0 10.81 8.95 2.73 1.78 0.00 0.00 0.00 0.00
12:16:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:16:01 AM eth0 10.81 8.83 2.69 1.78 0.00 0.00 0.00 0.00
12:17:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:17:01 AM eth0 10.90 8.95 2.69 1.78 0.00 0.00 0.00 0.00
12:18:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:18:01 AM eth0 11.46 9.07 2.90 1.85 0.00 0.00 0.00 0.00
12:19:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:19:01 AM eth0 13.18 10.80 3.36 2.16 0.00 0.00 0.00 0.00
12:20:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:20:01 AM eth0 10.73 8.72 2.69 1.59 0.00 0.00 0.00 0.00
12:21:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:21:01 AM eth0 0.50 0.00 0.03 0.00 0.00 0.00 0.00 0.00
12:22:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:22:01 AM eth0 0.72 0.22 0.07 0.03 0.00 0.00 0.00 0.00
12:23:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:23:01 AM eth0 0.95 0.10 0.17 0.06 0.00 0.00 0.00 0.00
12:24:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:24:01 AM eth0 0.77 0.02 0.07 0.00 0.00 0.00 0.00 0.00
12:25:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:25:01 AM eth0 2.05 0.62 0.70 0.45 0.00 0.00 0.00 0.00
12:26:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:26:01 AM eth0 0.57 0.07 0.03 0.00 0.00 0.00 0.00 0.00
12:27:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:27:01 AM eth0 68.08 66.48 18.26 7.48 0.00 0.00 0.00 0.00
12:28:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:28:01 AM eth0 153.34 107.33 434.04 295.05 0.00 0.00 0.00 0.00
12:29:01 AM lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
12:29:01 AM eth0 17.88 15.36 15.00 7.46 0.00 0.00 0.00 0.00
Average: lo 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
Average: eth0 13.51 10.74 17.30 11.40 0.00 0.00 0.00 0.00
---> sar -P ALL:
Linux 4.18.0-553.5.1.el8.x86_64 (prd-centos8-robot-2c-8g-1207.novalocal) 10/07/2025 _x86_64_ (2 CPU)
12:00:01 AM CPU %user %nice %system %iowait %steal %idle
12:01:01 AM all 4.71 0.03 0.46 0.02 0.08 94.71
12:01:01 AM 0 3.00 0.00 0.35 0.03 0.08 96.53
12:01:01 AM 1 6.42 0.05 0.57 0.00 0.08 92.88
12:02:01 AM all 0.33 0.00 0.08 0.00 0.07 99.53
12:02:01 AM 0 0.08 0.00 0.10 0.00 0.08 99.73
12:02:01 AM 1 0.57 0.00 0.05 0.00 0.05 99.33
12:03:01 AM all 0.23 0.00 0.08 0.00 0.07 99.62
12:03:01 AM 0 0.12 0.00 0.08 0.00 0.07 99.73
12:03:01 AM 1 0.33 0.00 0.08 0.00 0.07 99.52
12:04:01 AM all 0.35 0.00 0.06 0.00 0.06 99.53
12:04:01 AM 0 0.23 0.00 0.05 0.00 0.05 99.67
12:04:01 AM 1 0.47 0.00 0.07 0.00 0.07 99.40
12:05:01 AM all 0.34 0.00 0.06 0.00 0.07 99.53
12:05:01 AM 0 0.13 0.00 0.07 0.00 0.07 99.73
12:05:01 AM 1 0.55 0.00 0.05 0.00 0.07 99.33
12:06:01 AM all 0.45 0.00 0.08 0.00 0.06 99.41
12:06:01 AM 0 0.32 0.00 0.10 0.00 0.05 99.53
12:06:01 AM 1 0.58 0.00 0.07 0.00 0.07 99.28
12:07:01 AM all 0.33 0.00 0.08 0.01 0.05 99.53
12:07:01 AM 0 0.07 0.00 0.07 0.02 0.03 99.82
12:07:01 AM 1 0.60 0.00 0.08 0.00 0.07 99.25
12:08:01 AM all 2.07 0.00 0.38 0.01 0.07 97.48
12:08:01 AM 0 1.54 0.00 0.40 0.00 0.08 97.98
12:08:01 AM 1 2.61 0.00 0.35 0.02 0.05 96.98
12:09:01 AM all 0.34 0.00 0.07 0.00 0.06 99.53
12:09:01 AM 0 0.62 0.00 0.08 0.00 0.08 99.21
12:09:01 AM 1 0.07 0.00 0.05 0.00 0.03 99.85
12:10:01 AM all 0.28 0.19 0.20 0.00 0.09 99.24
12:10:01 AM 0 0.20 0.37 0.22 0.00 0.07 99.13
12:10:01 AM 1 0.36 0.00 0.19 0.00 0.10 99.35
12:11:01 AM all 0.78 0.00 0.12 0.01 0.08 99.01
12:11:01 AM 0 0.60 0.00 0.10 0.02 0.07 99.22
12:11:01 AM 1 0.95 0.00 0.13 0.00 0.10 98.81
12:11:01 AM CPU %user %nice %system %iowait %steal %idle
12:12:01 AM all 0.26 0.00 0.06 0.00 0.06 99.62
12:12:01 AM 0 0.42 0.00 0.05 0.00 0.07 99.46
12:12:01 AM 1 0.10 0.00 0.07 0.00 0.05 99.78
12:13:01 AM all 2.75 0.00 0.24 0.00 0.07 96.94
12:13:01 AM 0 3.21 0.00 0.27 0.00 0.07 96.45
12:13:01 AM 1 2.29 0.00 0.22 0.00 0.07 97.43
12:14:01 AM all 4.43 0.00 0.33 0.00 0.10 95.13
12:14:01 AM 0 1.79 0.00 0.20 0.00 0.08 97.93
12:14:01 AM 1 7.07 0.00 0.47 0.00 0.12 92.34
12:15:01 AM all 4.45 0.00 0.33 0.00 0.10 95.12
12:15:01 AM 0 2.51 0.00 0.22 0.00 0.08 97.19
12:15:01 AM 1 6.39 0.00 0.43 0.00 0.12 93.06
12:16:01 AM all 4.31 0.00 0.33 0.00 0.11 95.25
12:16:01 AM 0 3.00 0.00 0.27 0.00 0.10 96.63
12:16:01 AM 1 5.62 0.00 0.40 0.00 0.12 93.86
12:17:01 AM all 4.37 0.00 0.33 0.00 0.13 95.17
12:17:01 AM 0 1.30 0.00 0.18 0.00 0.10 98.41
12:17:01 AM 1 7.44 0.00 0.48 0.00 0.15 91.93
12:18:01 AM all 4.46 0.00 0.32 0.00 0.09 95.13
12:18:01 AM 0 6.54 0.00 0.44 0.00 0.10 92.92
12:18:01 AM 1 2.39 0.00 0.20 0.00 0.08 97.33
12:19:01 AM all 5.10 0.00 0.33 0.00 0.10 94.46
12:19:01 AM 0 3.56 0.00 0.30 0.00 0.10 96.04
12:19:01 AM 1 6.64 0.00 0.37 0.00 0.10 92.89
12:20:01 AM all 3.42 0.00 0.24 0.00 0.08 96.26
12:20:01 AM 0 3.06 0.00 0.25 0.00 0.10 96.59
12:20:01 AM 1 3.77 0.00 0.23 0.00 0.07 95.93
12:21:01 AM all 0.27 0.00 0.08 0.00 0.05 99.61
12:21:01 AM 0 0.12 0.00 0.08 0.00 0.07 99.73
12:21:01 AM 1 0.42 0.00 0.07 0.00 0.03 99.48
12:22:01 AM all 0.33 0.00 0.09 0.00 0.06 99.52
12:22:01 AM 0 0.15 0.00 0.12 0.00 0.08 99.65
12:22:01 AM 1 0.50 0.00 0.07 0.00 0.03 99.40
12:22:01 AM CPU %user %nice %system %iowait %steal %idle
12:23:01 AM all 0.28 0.00 0.07 0.00 0.06 99.60
12:23:01 AM 0 0.10 0.00 0.08 0.00 0.07 99.75
12:23:01 AM 1 0.45 0.00 0.05 0.00 0.05 99.45
12:24:01 AM all 0.34 0.00 0.08 0.00 0.07 99.52
12:24:01 AM 0 0.08 0.00 0.08 0.00 0.08 99.75
12:24:01 AM 1 0.60 0.00 0.07 0.00 0.05 99.28
12:25:01 AM all 0.45 0.00 0.08 0.00 0.05 99.42
12:25:01 AM 0 0.28 0.00 0.10 0.00 0.05 99.57
12:25:01 AM 1 0.62 0.00 0.07 0.00 0.05 99.27
12:26:01 AM all 0.32 0.00 0.06 0.00 0.07 99.56
12:26:01 AM 0 0.12 0.00 0.08 0.00 0.08 99.72
12:26:01 AM 1 0.52 0.00 0.03 0.00 0.05 99.40
12:27:01 AM all 8.21 0.00 1.36 0.00 0.07 90.37
12:27:01 AM 0 5.92 0.00 1.64 0.00 0.07 92.37
12:27:01 AM 1 10.50 0.00 1.07 0.00 0.07 88.36
12:28:01 AM all 24.78 0.00 2.68 0.38 0.10 72.07
12:28:01 AM 0 10.35 0.00 2.41 0.42 0.12 86.71
12:28:01 AM 1 39.20 0.00 2.94 0.33 0.08 57.44
12:29:01 AM all 28.92 0.00 4.00 0.14 0.11 66.84
12:29:01 AM 0 22.81 0.00 3.15 0.13 0.12 73.78
12:29:01 AM 1 35.02 0.00 4.84 0.15 0.10 59.89
Average: all 3.71 0.01 0.44 0.02 0.08 95.75
Average: 0 2.49 0.01 0.40 0.02 0.08 97.00
Average: 1 4.93 0.00 0.47 0.02 0.07 94.50