[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@67b9aa03}
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4e36e7ec}
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@5e5800bb}
[main] INFO org.opendaylight.transportpce.common.network.RequestProcessor - RequestProcessor instantiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id PceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2d4e55ba, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = PceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2d4e55ba, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=PceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728155280818
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandlerOperations with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2d4e55ba, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandlerOperations
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2d4e55ba, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandlerOperations] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728155280891
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandler with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2d4e55ba, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandler
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2d4e55ba, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandler] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728155280903
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id RendererListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2d4e55ba, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = RendererListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2d4e55ba, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=RendererListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728155280915
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding alarm topic: ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2b74db41, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@2b74db41, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728155280927
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - tapi converter: org.opendaylight.transportpce.common.converter.JsonStringConverter@46a8d91a
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - NbiNotificationsProvider Session Initiated
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - SendEvent request to topic 'Service' 
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Serialized event {"nbi-notifications:notification-process-service":{"connection-type":"service","service-z-end":{"clli":"clli","service-rate":1,"service-format":"OC","rx-direction":[{"index":0,"port":{"port-shelf":"port shelf","port-rack":"port rack","port-slot":"port slot","port-device-name":"device name","port-name":"port name","port-type":"port type","port-sub-slot":"port subslot"},"lgx":{"lgx-port-name":"lgx port name","lgx-port-shelf":"lgx port shelf","lgx-port-rack":"lgx port rack","lgx-device-name":"lgx device name"}}],"node-id":"XPONDER-1-2","is-split-lambda":false,"tx-direction":[{"index":0,"port":{"port-shelf":"port shelf","port-rack":"port rack","port-slot":"port slot","port-device-name":"device name","port-name":"port name","port-type":"port type","port-sub-slot":"port subslot"},"lgx":{"lgx-port-name":"lgx port name","lgx-port-shelf":"lgx port shelf","lgx-port-rack":"lgx port rack","lgx-device-name":"lgx device name"}}]},"common-id":"commond-id","service-a-end":{"clli":"clli","service-rate":1,"service-format":"OC","rx-direction":[{"index":0,"port":{"port-shelf":"port shelf","port-rack":"port rack","port-slot":"port slot","port-device-name":"device name","port-name":"port name","port-type":"port type","port-sub-slot":"port subslot"},"lgx":{"lgx-port-name":"lgx port name","lgx-port-shelf":"lgx port shelf","lgx-port-rack":"lgx port rack","lgx-device-name":"lgx device name"}}],"node-id":"XPONDER-1-2","is-split-lambda":false,"tx-direction":[{"index":0,"port":{"port-shelf":"port shelf","port-rack":"port rack","port-slot":"port slot","port-device-name":"device name","port-name":"port name","port-type":"port type","port-sub-slot":"port subslot"},"lgx":{"lgx-port-name":"lgx port name","lgx-port-shelf":"lgx port shelf","lgx-port-rack":"lgx port rack","lgx-device-name":"lgx device name"}}]},"service-name":"service1","aToZ":{"frequency":"196.125","rx-estimated-gsnr":"23.5","width":"87.5","optical-operational-mode":"test","max-output-power":"0.5","min-output-power":"0.5","rx-estimated-osnr":"23.5"},"operational-state":"inService","is-temp-service":true,"response-failed":"","zToA":{"frequency":"196.125","rx-estimated-gsnr":"23.5","width":"87.5","optical-operational-mode":"test","max-output-power":"0.5","min-output-power":"0.5","rx-estimated-osnr":"23.5"},"message":"message"}}
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@59bf89f2}
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@79a14807}
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@eba6581}
[main] INFO org.opendaylight.transportpce.common.network.RequestProcessor - RequestProcessor instantiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id PceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4f4a00a2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = PceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4f4a00a2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=PceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [PceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandlerOperations with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4f4a00a2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandlerOperations
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4f4a00a2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandlerOperations] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandlerOperations] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandler with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4f4a00a2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandler
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4f4a00a2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandler] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandler] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id RendererListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4f4a00a2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = RendererListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@4f4a00a2, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=RendererListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [RendererListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding alarm topic: ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@696c634d, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@696c634d, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - tapi converter: org.opendaylight.transportpce.common.converter.JsonStringConverter@29b421b
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - NbiNotificationsProvider Session Initiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.rpc.CreateNotificationSubscriptionServiceImpl - Adding T-API topic: 76d8f07b-ead5-4132-8eb8-cf3fdef7e079 to Kafka server
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding new tapi topic: 76d8f07b-ead5-4132-8eb8-cf3fdef7e079
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id 76d8f07b-ead5-4132-8eb8-cf3fdef7e079 with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@29b421b, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=76d8f07b-ead5-4132-8eb8-cf3fdef7e079}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = 76d8f07b-ead5-4132-8eb8-cf3fdef7e079
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@29b421b, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=76d8f07b-ead5-4132-8eb8-cf3fdef7e079}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=76d8f07b-ead5-4132-8eb8-cf3fdef7e079] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 3.7.0
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: 2ae524ed625438c5
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1728155281679
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Getting tapi notification context
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - SendEvent request to topic '' 
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer - Serialized event {"nbi-notifications:notification-tapi-service":{"layer-protocol-name":"ETH","changed-attributes":[{"value-name":"administrativeState","old-value":"LOCKED","new-value":"UNLOCKED"},{"value-name":"operationalState","old-value":"DISABLED","new-value":"ENABLED"}],"event-time-stamp":"2022-04-06T09:06:01+00:00","target-object-type":"tapi-connectivity:CONNECTIVITY_OBJECT_TYPE_CONNECTIVITY_SERVICE","notification-type":"tapi-notification:NOTIFICATION_TYPE_ATTRIBUTE_VALUE_CHANGE","target-object-identifier":"76d8f07b-ead5-4132-8eb8-cf3fdef7e079","target-object-name":[{"value-name":"Connectivity Service Name","value":"76d8f07b-ead5-4132-8eb8-cf3fdef7e079"}],"uuid":"6ae49721-7f96-4924-b23b-31e5bee4946b"}}
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@3c5c9798}
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@472dc7eb}
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.TapiNotificationSerializer - Deserializer configuration {converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@28c60940}
[main] INFO org.opendaylight.transportpce.common.network.RequestProcessor - RequestProcessor instantiated
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: PceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id PceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@37ed7dcd, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = PceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@37ed7dcd, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=PceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=PceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [PceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandlerOperations
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandlerOperations with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@37ed7dcd, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandlerOperations
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@37ed7dcd, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandlerOperations}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandlerOperations] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandlerOperations] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: ServiceHandler
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceHandler with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@37ed7dcd, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceHandler
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@37ed7dcd, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceHandler}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceHandler] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceHandler] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding process topic: RendererListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id RendererListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@37ed7dcd, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = RendererListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@37ed7dcd, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=RendererListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=RendererListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [RendererListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - Creating publisher for the following class ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.utils.TopicManager - Adding alarm topic: ServiceListener
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - Creation publisher for id ServiceListener with properties {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@aa8c70f, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] WARN org.apache.kafka.clients.CommonClientConfigs - Disabling exponential reconnect backoff because reconnect.backoff.ms is set, but reconnect.backoff.max.ms is not.
[main] INFO org.apache.kafka.clients.producer.ProducerConfig - ProducerConfig values: 
	acks = -1
	auto.include.jmx.reporter = true
	batch.size = 16384
	bootstrap.servers = [localhost:8080]
	buffer.memory = 33554432
	client.dns.lookup = use_all_dns_ips
	client.id = ServiceListener
	compression.type = none
	connections.max.idle.ms = 540000
	delivery.timeout.ms = 120000
	enable.idempotence = true
	enable.metrics.push = true
	interceptor.classes = []
	key.serializer = class org.apache.kafka.common.serialization.StringSerializer
	linger.ms = 1
	max.block.ms = 60000
	max.in.flight.requests.per.connection = 1
	max.request.size = 1048576
	metadata.max.age.ms = 300000
	metadata.max.idle.ms = 300000
	metric.reporters = []
	metrics.num.samples = 2
	metrics.recording.level = INFO
	metrics.sample.window.ms = 30000
	partitioner.adaptive.partitioning.enable = true
	partitioner.availability.timeout.ms = 0
	partitioner.class = null
	partitioner.ignore.keys = false
	receive.buffer.bytes = 32768
	reconnect.backoff.max.ms = 600000
	reconnect.backoff.ms = 600000
	request.timeout.ms = 30000
	retries = 3
	retry.backoff.max.ms = 1000
	retry.backoff.ms = 100
	sasl.client.callback.handler.class = null
	sasl.jaas.config = null
	sasl.kerberos.kinit.cmd = /usr/bin/kinit
	sasl.kerberos.min.time.before.relogin = 60000
	sasl.kerberos.service.name = null
	sasl.kerberos.ticket.renew.jitter = 0.05
	sasl.kerberos.ticket.renew.window.factor = 0.8
	sasl.login.callback.handler.class = null
	sasl.login.class = null
	sasl.login.connect.timeout.ms = null
	sasl.login.read.timeout.ms = null
	sasl.login.refresh.buffer.seconds = 300
	sasl.login.refresh.min.period.seconds = 60
	sasl.login.refresh.window.factor = 0.8
	sasl.login.refresh.window.jitter = 0.05
	sasl.login.retry.backoff.max.ms = 10000
	sasl.login.retry.backoff.ms = 100
	sasl.mechanism = GSSAPI
	sasl.oauthbearer.clock.skew.seconds = 30
	sasl.oauthbearer.expected.audience = null
	sasl.oauthbearer.expected.issuer = null
	sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000
	sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100
	sasl.oauthbearer.jwks.endpoint.url = null
	sasl.oauthbearer.scope.claim.name = scope
	sasl.oauthbearer.sub.claim.name = sub
	sasl.oauthbearer.token.endpoint.url = null
	security.protocol = PLAINTEXT
	security.providers = null
	send.buffer.bytes = 131072
	socket.connection.setup.timeout.max.ms = 30000
	socket.connection.setup.timeout.ms = 10000
	ssl.cipher.suites = null
	ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
	ssl.endpoint.identification.algorithm = https
	ssl.engine.factory.class = null
	ssl.key.password = null
	ssl.keymanager.algorithm = SunX509
	ssl.keystore.certificate.chain = null
	ssl.keystore.key = null
	ssl.keystore.location = null
	ssl.keystore.password = null
	ssl.keystore.type = JKS
	ssl.protocol = TLSv1.3
	ssl.provider = null
	ssl.secure.random.implementation = null
	ssl.trustmanager.algorithm = PKIX
	ssl.truststore.certificates = null
	ssl.truststore.location = null
	ssl.truststore.password = null
	ssl.truststore.type = JKS
	transaction.timeout.ms = 60000
	transactional.id = null
	value.serializer = class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer

[main] INFO org.apache.kafka.common.telemetry.internals.KafkaMetricsCollector - initializing Kafka metrics collector
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Deserializer configuration {acks=all, batch.size=16384, reconnect.backoff.ms=600000, converter=org.opendaylight.transportpce.common.converter.JsonStringConverter@aa8c70f, bootstrap.servers=localhost:8080, buffer.memory=33554432, key.serializer=class org.apache.kafka.common.serialization.StringSerializer, retries=3, value.serializer=class org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer, max.in.flight.requests.per.connection=1, linger.ms=1, client.id=ServiceListener}
[main] INFO org.apache.kafka.clients.producer.KafkaProducer - [Producer clientId=ServiceListener] Instantiated an idempotent producer.
[main] INFO org.apache.kafka.common.utils.AppInfoParser - The mbean of App info: [kafka.producer], id: [ServiceListener] already exists, so skipping a new mbean creation.
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - tapi converter: org.opendaylight.transportpce.common.converter.JsonStringConverter@5f181b9b
[main] INFO org.opendaylight.transportpce.nbinotifications.impl.NbiNotificationsProvider - NbiNotificationsProvider Session Initiated
[main] INFO org.opendaylight.transportpce.nbinotifications.producer.Publisher - SendEvent request to topic 'alarmservice' 
[main] INFO org.opendaylight.transportpce.nbinotifications.serialization.NotificationAlarmServiceSerializer - Serialized event {"nbi-notifications:notification-alarm-service":{"service-name":"service1","operational-state":"inService","connection-type":"service","message":"The service is now inService"}}