[main] INFO org.opendaylight.transportpce.common.network.RequestProcessor - RequestProcessor instantiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id PceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4434881a, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = PceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4434881a, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=PceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [PceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandlerOperations with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4434881a, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandlerOperations
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4434881a, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandlerOperations] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandlerOperations] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandler with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4434881a, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandler
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4434881a, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandler] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceHandlerOperations] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[kafka-producer-network-thread | PceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Node -1 disconnected.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandler] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id RendererListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4434881a, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = RendererListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4434881a, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=RendererListener] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceHandler] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [RendererListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding alarm topic: ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@50cf0f4f, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@50cf0f4f, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - tapi converter: org.opendaylight.transportpce.common.converter.JsonStringConverter@2b79ad77
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - NbiNotificationsProvider Session Initiated
[kafka-producer-network-thread | RendererListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Node -1 disconnected.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.GetNotificationsAlarmServiceImpl - RPC getNotificationsAlarmService received
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Subscribing for group id groupId, client config id consumerId with properties {key.deserializer=class org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceDeserializer, enable.auto.commit=false, group.id=groupId, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@555e74c, bootstrap.servers=localhost:8080, auto.commit.interval.ms=1000, auto.offset.reset=earliest, client.id=consumerId}
[kafka-producer-network-thread | ServiceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Node -1 disconnected.
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.consumer.ConsumerConfig - ConsumerConfig values: 
	allow.auto.create.topics = true
	auto.commit.interval.ms = 1000
	auto.include.jmx.reporter = true
	auto.offset.reset = earliest
	bootstrap.servers = [localhost:8080]
	check.crcs = true
	client.dns.lookup = use_all_dns_ips
	client.id = consumerId
	client.rack = 
	connections.max.idle.ms = 540000
	default.api.timeout.ms = 60000
	enable.auto.commit = false
	enable.metrics.push = true
	exclude.internal.topics = true
	fetch.max.bytes = 52428800
	fetch.max.wait.ms = 500
	fetch.min.bytes = 1
	group.id = groupId
	group.instance.id = null
	group.protocol = classic
	group.remote.assignor = null
	heartbeat.interval.ms = 3000
	interceptor.classes = []
	internal.leave.group.on.close = true
	internal.throw.on.fetch.stable.offset.unsupported = false
	isolation.level = read_uncommitted
	key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
	max.partition.fetch.bytes = 1048576
	max.poll.interval.ms = 300000
	max.poll.records = 500
	metadata.max.age.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor]
	receive.buffer.bytes = 65536
	reconnect.backoff.max.ms = 1000
	reconnect.backoff.ms = 50
	request.timeout.ms = 30000
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	session.timeout.ms = 45000
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	value.deserializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceDeserializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceDeserializer - Deserializer configuration {key.deserializer=class org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceDeserializer, enable.auto.commit=false, group.id=groupId, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@555e74c, bootstrap.servers=localhost:8080, auto.commit.interval.ms=1000, auto.offset.reset=earliest, client.id=consumerId}
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728344302350
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Subscribe request to topic 'alarmservice' 
[main] INFO org.apache.kafka.clients.consumer.internals.LegacyKafkaConsumer - [Consumer clientId=consumerId, groupId=groupId] Subscribed to topic(s): alarmservice
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Getting records '[]' 
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=consumerId, groupId=groupId] Resetting generation and member id due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=consumerId, groupId=groupId] Request joining group due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.LegacyKafkaConsumer - [Consumer clientId=consumerId, groupId=groupId] Unsubscribed all topics or patterns and assigned partitions
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=consumerId, groupId=groupId] Resetting generation and member id due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=consumerId, groupId=groupId] Request joining group due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.common.metrics.Metrics - Metrics scheduler closed
[main] INFO org.apache.kafka.common.metrics.Metrics - Closing reporter org.apache.kafka.common.metrics.JmxReporter
[main] INFO org.apache.kafka.common.metrics.Metrics - Closing reporter org.apache.kafka.common.telemetry.internals.ClientTelemetryReporter
[main] INFO org.apache.kafka.common.metrics.Metrics - Metrics reporters closed
[main] INFO org.apache.kafka.common.utils.AppInfoParser - App info kafka.consumer for consumerId unregistered
[main] INFO org.opendaylight.transportpce.common.network.RequestProcessor - RequestProcessor instantiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id PceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2a1c965c, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = PceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2a1c965c, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=PceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [PceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandlerOperations with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2a1c965c, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandlerOperations
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2a1c965c, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandlerOperations] Instantiated an idempotent producer.
[kafka-producer-network-thread | PceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Node -1 disconnected.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandlerOperations] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandler with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2a1c965c, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandler
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2a1c965c, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandler] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceHandlerOperations] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandler] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id RendererListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2a1c965c, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = RendererListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2a1c965c, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=RendererListener] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceHandler] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [RendererListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding alarm topic: ServiceListener
[kafka-producer-network-thread | RendererListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Node -1 disconnected.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@7c381e2f, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@7c381e2f, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - tapi converter: org.opendaylight.transportpce.common.converter.JsonStringConverter@5aad3221
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - NbiNotificationsProvider Session Initiated
[kafka-producer-network-thread | ServiceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Node -1 disconnected.
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.GetNotificationsProcessServiceImpl - RPC getNotificationsService received
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Subscribing for group id groupId, client config id consumerId with properties {key.deserializer=class org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceDeserializer, enable.auto.commit=false, group.id=groupId, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@56419cf6, bootstrap.servers=localhost:8080, auto.commit.interval.ms=1000, auto.offset.reset=earliest, client.id=consumerId}
[main] INFO org.apache.kafka.clients.consumer.ConsumerConfig - ConsumerConfig values: 
	allow.auto.create.topics = true
	auto.commit.interval.ms = 1000
	auto.include.jmx.reporter = true
	auto.offset.reset = earliest
	bootstrap.servers = [localhost:8080]
	check.crcs = true
	client.dns.lookup = use_all_dns_ips
	client.id = consumerId
	client.rack = 
	connections.max.idle.ms = 540000
	default.api.timeout.ms = 60000
	enable.auto.commit = false
	enable.metrics.push = true
	exclude.internal.topics = true
	fetch.max.bytes = 52428800
	fetch.max.wait.ms = 500
	fetch.min.bytes = 1
	group.id = groupId
	group.instance.id = null
	group.protocol = classic
	group.remote.assignor = null
	heartbeat.interval.ms = 3000
	interceptor.classes = []
	internal.leave.group.on.close = true
	internal.throw.on.fetch.stable.offset.unsupported = false
	isolation.level = read_uncommitted
	key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
	max.partition.fetch.bytes = 1048576
	max.poll.interval.ms = 300000
	max.poll.records = 500
	metadata.max.age.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor]
	receive.buffer.bytes = 65536
	reconnect.backoff.max.ms = 1000
	reconnect.backoff.ms = 50
	request.timeout.ms = 30000
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	session.timeout.ms = 45000
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	value.deserializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceDeserializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceDeserializer - Deserializer configuration {key.deserializer=class org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceDeserializer, enable.auto.commit=false, group.id=groupId, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@56419cf6, bootstrap.servers=localhost:8080, auto.commit.interval.ms=1000, auto.offset.reset=earliest, client.id=consumerId}
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728344303437
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Subscribe request to topic 'service' 
[main] INFO org.apache.kafka.clients.consumer.internals.LegacyKafkaConsumer - [Consumer clientId=consumerId, groupId=groupId] Subscribed to topic(s): service
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=consumerId, groupId=groupId] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Getting records '[]' 
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=consumerId, groupId=groupId] Resetting generation and member id due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=consumerId, groupId=groupId] Request joining group due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.LegacyKafkaConsumer - [Consumer clientId=consumerId, groupId=groupId] Unsubscribed all topics or patterns and assigned partitions
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=consumerId, groupId=groupId] Resetting generation and member id due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=consumerId, groupId=groupId] Request joining group due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.common.metrics.Metrics - Metrics scheduler closed
[main] INFO org.apache.kafka.common.metrics.Metrics - Closing reporter org.apache.kafka.common.metrics.JmxReporter
[main] INFO org.apache.kafka.common.metrics.Metrics - Closing reporter org.apache.kafka.common.telemetry.internals.ClientTelemetryReporter
[main] INFO org.apache.kafka.common.metrics.Metrics - Metrics reporters closed
[main] INFO org.apache.kafka.common.utils.AppInfoParser - App info kafka.consumer for consumerId unregistered
[main] INFO org.opendaylight.transportpce.common.network.RequestProcessor - RequestProcessor instantiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id PceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2550cbc8, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = PceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2550cbc8, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=PceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [PceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandlerOperations with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2550cbc8, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandlerOperations
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[kafka-producer-network-thread | PceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Node -1 disconnected.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2550cbc8, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandlerOperations] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandlerOperations] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandler with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2550cbc8, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandler
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2550cbc8, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandler] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandler] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id RendererListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2550cbc8, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = RendererListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2550cbc8, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=RendererListener] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceHandlerOperations] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [RendererListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding alarm topic: ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@7ea2191d, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@7ea2191d, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceListener] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceHandler] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[kafka-producer-network-thread | RendererListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Node -1 disconnected.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - tapi converter: org.opendaylight.transportpce.common.converter.JsonStringConverter@370cf017
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - NbiNotificationsProvider Session Initiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.CreateNotificationSubscriptionServiceImpl - Adding T-API topic: 05515a93-3d3f-49ae-8874-24882bffe6b6 to Kafka server
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding new tapi topic: 05515a93-3d3f-49ae-8874-24882bffe6b6
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id 05515a93-3d3f-49ae-8874-24882bffe6b6 with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@370cf017, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=05515a93-3d3f-49ae-8874-24882bffe6b6}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = 05515a93-3d3f-49ae-8874-24882bffe6b6
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@370cf017, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=05515a93-3d3f-49ae-8874-24882bffe6b6}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=05515a93-3d3f-49ae-8874-24882bffe6b6] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Node -1 disconnected.
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728344304495
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Getting tapi notification context
[kafka-producer-network-thread | 05515a93-3d3f-49ae-8874-24882bffe6b6] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=05515a93-3d3f-49ae-8874-24882bffe6b6] Node -1 disconnected.
[kafka-producer-network-thread | 05515a93-3d3f-49ae-8874-24882bffe6b6] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=05515a93-3d3f-49ae-8874-24882bffe6b6] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | 05515a93-3d3f-49ae-8874-24882bffe6b6] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=05515a93-3d3f-49ae-8874-24882bffe6b6] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.common.network.RequestProcessor - RequestProcessor instantiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id PceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@1320e18e, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = PceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@1320e18e, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=PceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [PceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandlerOperations with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@1320e18e, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandlerOperations
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@1320e18e, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandlerOperations] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandlerOperations] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandler with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@1320e18e, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandler
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[kafka-producer-network-thread | PceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Node -1 disconnected.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@1320e18e, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandler] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandler] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id RendererListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@1320e18e, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = RendererListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@1320e18e, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=RendererListener] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceHandlerOperations] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandler] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [RendererListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding alarm topic: ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@5a0a1c75, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@5a0a1c75, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceListener] Instantiated an idempotent producer.
[kafka-producer-network-thread | RendererListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Node -1 disconnected.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - tapi converter: org.opendaylight.transportpce.common.converter.JsonStringConverter@3c8511c2
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - NbiNotificationsProvider Session Initiated
[kafka-producer-network-thread | ServiceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Node -1 disconnected.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.CreateNotificationSubscriptionServiceImpl - Adding T-API topic: 4edfe1d3-5f8f-4873-8766-cc45195d1c8c to Kafka server
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding new tapi topic: 4edfe1d3-5f8f-4873-8766-cc45195d1c8c
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id 4edfe1d3-5f8f-4873-8766-cc45195d1c8c with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@3c8511c2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=4edfe1d3-5f8f-4873-8766-cc45195d1c8c}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = 4edfe1d3-5f8f-4873-8766-cc45195d1c8c
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@3c8511c2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=4edfe1d3-5f8f-4873-8766-cc45195d1c8c}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728344304539
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Getting tapi notification context
[kafka-producer-network-thread | 4edfe1d3-5f8f-4873-8766-cc45195d1c8c] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Node -1 disconnected.
[kafka-producer-network-thread | 4edfe1d3-5f8f-4873-8766-cc45195d1c8c] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | 4edfe1d3-5f8f-4873-8766-cc45195d1c8c] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.GetNotificationListImpl - RPC getNotificationList received
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.GetNotificationListImpl - Going to get notifications for topic 4edfe1d3-5f8f-4873-8766-cc45195d1c8c
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Subscribing for group id 4edfe1d3-5f8f-4873-8766-cc45195d1c8c, client config id 4edfe1d3-5f8f-4873-8766-cc45195d1c8c with properties {key.deserializer=class org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationDeserializer, enable.auto.commit=false, group.id=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@434e8cdb, bootstrap.servers=localhost:8080, auto.commit.interval.ms=1000, auto.offset.reset=earliest, client.id=4edfe1d3-5f8f-4873-8766-cc45195d1c8c}
[main] INFO org.apache.kafka.clients.consumer.ConsumerConfig - ConsumerConfig values: 
	allow.auto.create.topics = true
	auto.commit.interval.ms = 1000
	auto.include.jmx.reporter = true
	auto.offset.reset = earliest
	bootstrap.servers = [localhost:8080]
	check.crcs = true
	client.dns.lookup = use_all_dns_ips
	client.id = 4edfe1d3-5f8f-4873-8766-cc45195d1c8c
	client.rack = 
	connections.max.idle.ms = 540000
	default.api.timeout.ms = 60000
	enable.auto.commit = false
	enable.metrics.push = true
	exclude.internal.topics = true
	fetch.max.bytes = 52428800
	fetch.max.wait.ms = 500
	fetch.min.bytes = 1
	group.id = 4edfe1d3-5f8f-4873-8766-cc45195d1c8c
	group.instance.id = null
	group.protocol = classic
	group.remote.assignor = null
	heartbeat.interval.ms = 3000
	interceptor.classes = []
	internal.leave.group.on.close = true
	internal.throw.on.fetch.stable.offset.unsupported = false
	isolation.level = read_uncommitted
	key.deserializer = class org.apache.kafka.common.serialization.StringDeserializer
	max.partition.fetch.bytes = 1048576
	max.poll.interval.ms = 300000
	max.poll.records = 500
	metadata.max.age.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partition.assignment.strategy = [class org.apache.kafka.clients.consumer.RangeAssignor, class org.apache.kafka.clients.consumer.CooperativeStickyAssignor]
	receive.buffer.bytes = 65536
	reconnect.backoff.max.ms = 1000
	reconnect.backoff.ms = 50
	request.timeout.ms = 30000
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	session.timeout.ms = 45000
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	value.deserializer = class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationDeserializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationDeserializer - Tapi Deserializer configuration {key.deserializer=class org.apache.kafka.common.serialization.StringDeserializer, value.deserializer=class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationDeserializer, enable.auto.commit=false, group.id=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@434e8cdb, bootstrap.servers=localhost:8080, auto.commit.interval.ms=1000, auto.offset.reset=earliest, client.id=4edfe1d3-5f8f-4873-8766-cc45195d1c8c}
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728344304550
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Subscribe request to topic '4edfe1d3-5f8f-4873-8766-cc45195d1c8c' 
[main] INFO org.apache.kafka.clients.consumer.internals.LegacyKafkaConsumer - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Subscribed to topic(s): 4edfe1d3-5f8f-4873-8766-cc45195d1c8c
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Node -1 disconnected.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[main] WARN org.apache.kafka.clients.NetworkClient - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.opendaylight.transportpce.nbinotifications.consumer.Subscriber - Getting records '[]' 
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Resetting generation and member id due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Request joining group due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.LegacyKafkaConsumer - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Unsubscribed all topics or patterns and assigned partitions
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Resetting generation and member id due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.clients.consumer.internals.ConsumerCoordinator - [Consumer clientId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c, groupId=4edfe1d3-5f8f-4873-8766-cc45195d1c8c] Request joining group due to: consumer pro-actively leaving the group
[main] INFO org.apache.kafka.common.metrics.Metrics - Metrics scheduler closed
[main] INFO org.apache.kafka.common.metrics.Metrics - Closing reporter org.apache.kafka.common.metrics.JmxReporter
[main] INFO org.apache.kafka.common.metrics.Metrics - Closing reporter org.apache.kafka.common.telemetry.internals.ClientTelemetryReporter
[main] INFO org.apache.kafka.common.metrics.Metrics - Metrics reporters closed
[main] INFO org.apache.kafka.common.utils.AppInfoParser - App info kafka.consumer for 4edfe1d3-5f8f-4873-8766-cc45195d1c8c unregistered
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.GetNotificationListImpl - TAPI notifications = []
[main] INFO org.opendaylight.transportpce.common.network.RequestProcessor - RequestProcessor instantiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id PceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@32845349, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = PceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@32845349, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=PceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [PceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandlerOperations with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@32845349, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandlerOperations
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@32845349, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandlerOperations] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandlerOperations] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandler with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@32845349, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandler
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@32845349, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandler] Instantiated an idempotent producer.
[kafka-producer-network-thread | PceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Node -1 disconnected.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | PceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=PceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[kafka-producer-network-thread | ServiceHandlerOperations] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandlerOperations] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandlerOperations] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandler] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id RendererListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@32845349, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = RendererListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@32845349, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=RendererListener] Instantiated an idempotent producer.
[kafka-producer-network-thread | ServiceHandler] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Node -1 disconnected.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceHandler] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceHandler] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [RendererListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding alarm topic: ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@216e6013, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@216e6013, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceListener] Instantiated an idempotent producer.
[kafka-producer-network-thread | RendererListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Node -1 disconnected.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | RendererListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=RendererListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - tapi converter: org.opendaylight.transportpce.common.converter.JsonStringConverter@7f87098d
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - NbiNotificationsProvider Session Initiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.GetNotificationsProcessServiceImpl - RPC getNotificationsService received
[main] WARN org.opendaylight.transportpce.nbinotifications.impl.rpc.GetNotificationsProcessServiceImpl - Missing mandatory params for input GetNotificationsProcessServiceInput{}
[kafka-producer-network-thread | ServiceListener] INFO org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Node -1 disconnected.
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Connection to node -1 (localhost/127.0.0.1:8080) could not be established. Node may not be available.
[kafka-producer-network-thread | ServiceListener] WARN org.apache.kafka.clients.NetworkClient - [Producer clientId=ServiceListener] Bootstrap broker localhost:8080 (id: -1 rack: null) disconnected