timestamp,message 1649809721995,00:28:41 [] [] [] [] INFO ebRequestDiagnostics Handling GET /inventory/items/c4f68c59-ce41-59dd-8c7a-24fd725f5392 1649809721997,00:28:41 [] [] [] [] INFO ebRequestDiagnostics Handling GET /inventory/items/12b13a1a-7cec-5327-ac96-36eda10b509e 1649809722001,00:28:42 [] [] [] [] INFO ebRequestDiagnostics Handling GET /inventory/items/5e6897ff-c0c7-5286-832c-404eb44dc42c 1649809722017,00:28:42 [] [] [] [] INFO ebRequestDiagnostics Handling GET /inventory/items/37803b2f-ac73-5f6e-982d-023cb3c6c9eb 1649809722018,00:28:42 [] [] [] [] INFO ebRequestDiagnostics Handling GET /inventory/items/87dd891a-713d-54c4-8756-f1594da432c7 1649809722023,00:28:42 [] [] [] [] INFO ebRequestDiagnostics Handling GET /inventory/items/507774f2-7da0-5d11-af60-6f40d1326794 1649809722032,00:28:42 [] [] [] [] INFO ebRequestDiagnostics Handling GET /inventory/items/2eb35030-9891-56bf-a4df-d244ff0e8f5c 1649845482213,"10:24:42 [] [] [] [] INFO taImportKafkaHandler Data import event payload has been received with event type: DI_SRS_MARC_BIB_RECORD_CREATED, recordId: 6b698831-3ed8-4c66-87dc-5b5e21b4677a" 1649845482223,"10:24:42 [] [] [] [] INFO taImportKafkaHandler Data import event payload has been received with event type: DI_SRS_MARC_BIB_RECORD_CREATED, recordId: 96ef2306-67c7-4c53-bd48-6e73b7612686" 1649845482379,10:24:42 [] [] [] [] INFO ProfileSnapshotCache JobProfileSnapshot was loaded by id '9b098573-d8cc-4195-b621-e78e04563f3a' 1649845482498,10:24:42 [] [] [] [] INFO MappingMetadataCache MappingMetadata was loaded by jobExecutionId 'b2900193-00ea-465a-bd89-bece3690f701' 1649845482615,10:24:42 [] [] [] [] INFO AbstractConfig ProducerConfig values: 1649845482615, acks = -1 1649845482615, batch.size = 16384 1649845482615, bootstrap.servers = [kafka.kmsd.folio-eis.us-east-1:9092] 1649845482615, buffer.memory = 33554432 1649845482615, client.dns.lookup = default 1649845482615, client.id = producer-255 1649845482615, compression.type = gzip 1649845482615, connections.max.idle.ms = 540000 1649845482615, delivery.timeout.ms = 120000 1649845482615, enable.idempotence = true 1649845482615, interceptor.classes = [] 1649845482615, key.serializer = class org.apache.kafka.common.serialization.StringSerializer 1649845482615, linger.ms = 0 1649845482615, max.block.ms = 60000 1649845482615, max.in.flight.requests.per.connection = 5 1649845482615, max.request.size = 1048576 1649845482615, metadata.max.age.ms = 300000 1649845482615, metadata.max.idle.ms = 300000 1649845482615, metric.reporters = [] 1649845482615, metrics.num.samples = 2 1649845482615, metrics.recording.level = INFO 1649845482615, metrics.sample.window.ms = 30000 1649845482615, partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner 1649845482615, receive.buffer.bytes = 32768 1649845482615, reconnect.backoff.max.ms = 1000 1649845482615, reconnect.backoff.ms = 50 1649845482615, request.timeout.ms = 30000 1649845482615, retries = 2147483647 1649845482615, retry.backoff.ms = 100 1649845482615, sasl.client.callback.handler.class = null 1649845482615, sasl.jaas.config = null 1649845482615, sasl.kerberos.kinit.cmd = /usr/bin/kinit 1649845482615, sasl.kerberos.min.time.before.relogin = 60000 1649845482615, sasl.kerberos.service.name = null 1649845482615, sasl.kerberos.ticket.renew.jitter = 0.05 1649845482615, sasl.kerberos.ticket.renew.window.factor = 0.8 1649845482615, sasl.login.callback.handler.class = null 1649845482615, sasl.login.class = null 1649845482615, sasl.login.refresh.buffer.seconds = 300 1649845482615, sasl.login.refresh.min.period.seconds = 60 1649845482615, sasl.login.refresh.window.factor = 0.8 1649845482615, sasl.login.refresh.window.jitter = 0.05 1649845482615, sasl.mechanism = GSSAPI 1649845482615, security.protocol = PLAINTEXT 1649845482615, security.providers = null 1649845482615, send.buffer.bytes = 131072 1649845482615, ssl.cipher.suites = null 1649845482615, ssl.enabled.protocols = [TLSv1.2] 1649845482615, ssl.endpoint.identification.algorithm = null 1649845482615, ssl.key.password = null 1649845482615, ssl.keymanager.algorithm = SunX509 1649845482615, ssl.keystore.location = null 1649845482615, ssl.keystore.password = null 1649845482615, ssl.keystore.type = JKS 1649845482615, ssl.protocol = TLSv1.2 1649845482615, ssl.provider = null 1649845482615, ssl.secure.random.implementation = null 1649845482615, ssl.trustmanager.algorithm = PKIX 1649845482615, ssl.truststore.location = null 1649845482615, ssl.truststore.password = null 1649845482615, ssl.truststore.type = JKS 1649845482615, transaction.timeout.ms = 60000 1649845482615, transactional.id = null 1649845482615, value.serializer = class org.apache.kafka.common.serialization.StringSerializer 1649845482615,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-255] Instantiated an idempotent producer. 1649845482616,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-255] Overriding the default retries config to the recommended value of 2147483647 since the idempotent producer is enabled. 1649845482616,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-255] Overriding the default acks to all since idempotence is enabled. 1649845482617,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka version: 2.5.0 1649845482617,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka commitId: 66563e712b0b9f84 1649845482617,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka startTimeMs: 1649845482617 1649845482619,10:24:42 [] [] [] [] INFO Metadata [Producer clientId=producer-255] Cluster ID: G1_s3A4NTtuR07X2sF0UCg 1649845482619,10:24:42 [] [] [] [] INFO TransactionManager [Producer clientId=producer-255] ProducerId set to 148842 with epoch 0 1649845482627,10:24:42 [] [] [] [] INFO KafkaEventPublisher Event with type: DI_INVENTORY_INSTANCE_MATCHED and recordId: 6b698831-3ed8-4c66-87dc-5b5e21b4677a was sent to the topic kmsd.Default.fs00001019.DI_INVENTORY_INSTANCE_MATCHED 1649845482627,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-255] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. 1649845482630,"10:24:42 [] [] [] [] INFO taImportKafkaHandler Data import event payload has been received with event type: DI_INVENTORY_INSTANCE_MATCHED, recordId: 6b698831-3ed8-4c66-87dc-5b5e21b4677a" 1649845482653,10:24:42 [] [] [] [] INFO AbstractConfig ProducerConfig values: 1649845482653, acks = -1 1649845482653, batch.size = 16384 1649845482653, bootstrap.servers = [kafka.kmsd.folio-eis.us-east-1:9092] 1649845482653, buffer.memory = 33554432 1649845482653, client.dns.lookup = default 1649845482653, client.id = producer-256 1649845482653, compression.type = gzip 1649845482653, connections.max.idle.ms = 540000 1649845482653, delivery.timeout.ms = 120000 1649845482653, enable.idempotence = true 1649845482653, interceptor.classes = [] 1649845482653, key.serializer = class org.apache.kafka.common.serialization.StringSerializer 1649845482653, linger.ms = 0 1649845482653, max.block.ms = 60000 1649845482653, max.in.flight.requests.per.connection = 5 1649845482653, max.request.size = 1048576 1649845482653, metadata.max.age.ms = 300000 1649845482653, metadata.max.idle.ms = 300000 1649845482653, metric.reporters = [] 1649845482653, metrics.num.samples = 2 1649845482653, metrics.recording.level = INFO 1649845482653, metrics.sample.window.ms = 30000 1649845482653, partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner 1649845482653, receive.buffer.bytes = 32768 1649845482653, reconnect.backoff.max.ms = 1000 1649845482653, reconnect.backoff.ms = 50 1649845482653, request.timeout.ms = 30000 1649845482653, retries = 2147483647 1649845482653, retry.backoff.ms = 100 1649845482653, sasl.client.callback.handler.class = null 1649845482653, sasl.jaas.config = null 1649845482653, sasl.kerberos.kinit.cmd = /usr/bin/kinit 1649845482653, sasl.kerberos.min.time.before.relogin = 60000 1649845482653, sasl.kerberos.service.name = null 1649845482653, sasl.kerberos.ticket.renew.jitter = 0.05 1649845482653, sasl.kerberos.ticket.renew.window.factor = 0.8 1649845482653, sasl.login.callback.handler.class = null 1649845482653, sasl.login.class = null 1649845482653, sasl.login.refresh.buffer.seconds = 300 1649845482653, sasl.login.refresh.min.period.seconds = 60 1649845482653, sasl.login.refresh.window.factor = 0.8 1649845482653, sasl.login.refresh.window.jitter = 0.05 1649845482653, sasl.mechanism = GSSAPI 1649845482653, security.protocol = PLAINTEXT 1649845482653, security.providers = null 1649845482653, send.buffer.bytes = 131072 1649845482653, ssl.cipher.suites = null 1649845482653, ssl.enabled.protocols = [TLSv1.2] 1649845482653, ssl.endpoint.identification.algorithm = null 1649845482653, ssl.key.password = null 1649845482653, ssl.keymanager.algorithm = SunX509 1649845482653, ssl.keystore.location = null 1649845482653, ssl.keystore.password = null 1649845482653, ssl.keystore.type = JKS 1649845482653, ssl.protocol = TLSv1.2 1649845482653, ssl.provider = null 1649845482653, ssl.secure.random.implementation = null 1649845482653, ssl.trustmanager.algorithm = PKIX 1649845482653, ssl.truststore.location = null 1649845482653, ssl.truststore.password = null 1649845482653, ssl.truststore.type = JKS 1649845482653, transaction.timeout.ms = 60000 1649845482653, transactional.id = null 1649845482653, value.serializer = class org.apache.kafka.common.serialization.StringSerializer 1649845482653,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-256] Instantiated an idempotent producer. 1649845482654,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-256] Overriding the default retries config to the recommended value of 2147483647 since the idempotent producer is enabled. 1649845482654,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-256] Overriding the default acks to all since idempotence is enabled. 1649845482654,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka version: 2.5.0 1649845482654,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka commitId: 66563e712b0b9f84 1649845482654,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka startTimeMs: 1649845482654 1649845482657,10:24:42 [] [] [] [] INFO Metadata [Producer clientId=producer-256] Cluster ID: G1_s3A4NTtuR07X2sF0UCg 1649845482657,10:24:42 [] [] [] [] INFO TransactionManager [Producer clientId=producer-256] ProducerId set to 148843 with epoch 0 1649845482664,10:24:42 [] [] [] [] INFO KafkaEventPublisher Event with type: DI_INVENTORY_INSTANCE_MATCHED and recordId: 96ef2306-67c7-4c53-bd48-6e73b7612686 was sent to the topic kmsd.Default.fs00001019.DI_INVENTORY_INSTANCE_MATCHED 1649845482664,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-256] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. 1649845482667,"10:24:42 [] [] [] [] INFO taImportKafkaHandler Data import event payload has been received with event type: DI_INVENTORY_INSTANCE_MATCHED, recordId: 96ef2306-67c7-4c53-bd48-6e73b7612686" 1649845482670,10:24:42 [] [] [] [] INFO MappingMetadataCache MappingMetadata was loaded by jobExecutionId 'b2900193-00ea-465a-bd89-bece3690f701' 1649845482680,10:24:42 [] [] [] [] ERROR InstanceEventHandler Error updating inventory Instance 1649845482680,java.lang.NumberFormatException: null 1649845482680, at java.lang.Integer.parseInt(Integer.java:614) ~[?:?] 1649845482680, at java.lang.Integer.parseInt(Integer.java:770) ~[?:?] 1649845482680, at org.folio.inventory.support.InstanceUtil.mergeFieldsWhichAreNotControlled(InstanceUtil.java:44) ~[ms.jar:?] 1649845482680, at org.folio.inventory.dataimport.handlers.actions.ReplaceInstanceEventHandler.prepareAndExecuteMapping(ReplaceInstanceEventHandler.java:170) ~[ms.jar:?] 1649845482680, at org.folio.inventory.dataimport.handlers.actions.ReplaceInstanceEventHandler.lambda$handle$0(ReplaceInstanceEventHandler.java:83) ~[ms.jar:?] 1649845482680, at java.util.Optional.map(Optional.java:265) ~[?:?] 1649845482680, at org.folio.inventory.dataimport.handlers.actions.ReplaceInstanceEventHandler.lambda$handle$2(ReplaceInstanceEventHandler.java:83) ~[ms.jar:?] 1649845482680, at io.vertx.core.impl.future.Composition.onSuccess(Composition.java:38) ~[ms.jar:?] 1649845482680, at io.vertx.core.impl.future.FutureBase.emitSuccess(FutureBase.java:62) ~[ms.jar:?] 1649845482680, at io.vertx.core.impl.future.FutureImpl.tryComplete(FutureImpl.java:179) ~[ms.jar:?] 1649845482680, at io.vertx.core.impl.future.PromiseImpl.tryComplete(PromiseImpl.java:23) ~[ms.jar:?] 1649845482680, at io.vertx.core.Promise.complete(Promise.java:66) ~[ms.jar:?] 1649845482680, at io.vertx.core.Future.lambda$fromCompletionStage$3(Future.java:384) ~[ms.jar:?] 1649845482680, at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859) ~[?:?] 1649845482680, at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?] 1649845482680, at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?] 1649845482680, at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:2073) ~[?:?] 1649845482680, at io.vertx.ext.web.client.impl.HttpContext.handleDispatchResponse(HttpContext.java:369) ~[ms.jar:?] 1649845482680, at io.vertx.ext.web.client.impl.HttpContext.execute(HttpContext.java:356) ~[ms.jar:?] 1649845482680, at io.vertx.ext.web.client.impl.HttpContext.next(HttpContext.java:334) ~[ms.jar:?] 1649845482680, at io.vertx.ext.web.client.impl.HttpContext.fire(HttpContext.java:301) ~[ms.jar:?] 1649845482680, at io.vertx.ext.web.client.impl.HttpContext.dispatchResponse(HttpContext.java:263) ~[ms.jar:?] 1649845482680, at io.vertx.ext.web.client.impl.HttpContext.lambda$null$8(HttpContext.java:533) ~[ms.jar:?] 1649845482680, at io.vertx.core.impl.AbstractContext.dispatch(AbstractContext.java:96) ~[ms.jar:?] 1649845482680, at io.vertx.core.impl.WorkerContext.lambda$run$1(WorkerContext.java:102) ~[ms.jar:?] 1649845482680, at io.vertx.core.impl.TaskQueue.run(TaskQueue.java:76) ~[ms.jar:?] 1649845482680, at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] 1649845482680, at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] 1649845482680, at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) [ms.jar:?] 1649845482680, at java.lang.Thread.run(Thread.java:829) [?:?] 1649845482680,10:24:42 [] [] [] [] INFO AbstractConfig ProducerConfig values: 1649845482680, acks = -1 1649845482680, batch.size = 16384 1649845482680, bootstrap.servers = [kafka.kmsd.folio-eis.us-east-1:9092] 1649845482680, buffer.memory = 33554432 1649845482680, client.dns.lookup = default 1649845482680, client.id = producer-257 1649845482680, compression.type = gzip 1649845482680, connections.max.idle.ms = 540000 1649845482680, delivery.timeout.ms = 120000 1649845482680, enable.idempotence = true 1649845482680, interceptor.classes = [] 1649845482680, key.serializer = class org.apache.kafka.common.serialization.StringSerializer 1649845482680, linger.ms = 0 1649845482680, max.block.ms = 60000 1649845482680, max.in.flight.requests.per.connection = 5 1649845482680, max.request.size = 1048576 1649845482680, metadata.max.age.ms = 300000 1649845482680, metadata.max.idle.ms = 300000 1649845482680, metric.reporters = [] 1649845482680, metrics.num.samples = 2 1649845482680, metrics.recording.level = INFO 1649845482680, metrics.sample.window.ms = 30000 1649845482680, partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner 1649845482680, receive.buffer.bytes = 32768 1649845482680, reconnect.backoff.max.ms = 1000 1649845482680, reconnect.backoff.ms = 50 1649845482680, request.timeout.ms = 30000 1649845482680, retries = 2147483647 1649845482680, retry.backoff.ms = 100 1649845482680, sasl.client.callback.handler.class = null 1649845482680, sasl.jaas.config = null 1649845482680, sasl.kerberos.kinit.cmd = /usr/bin/kinit 1649845482680, sasl.kerberos.min.time.before.relogin = 60000 1649845482680, sasl.kerberos.service.name = null 1649845482680, sasl.kerberos.ticket.renew.jitter = 0.05 1649845482680, sasl.kerberos.ticket.renew.window.factor = 0.8 1649845482680, sasl.login.callback.handler.class = null 1649845482680, sasl.login.class = null 1649845482680, sasl.login.refresh.buffer.seconds = 300 1649845482680, sasl.login.refresh.min.period.seconds = 60 1649845482680, sasl.login.refresh.window.factor = 0.8 1649845482680, sasl.login.refresh.window.jitter = 0.05 1649845482680, sasl.mechanism = GSSAPI 1649845482680, security.protocol = PLAINTEXT 1649845482680, security.providers = null 1649845482680, send.buffer.bytes = 131072 1649845482680, ssl.cipher.suites = null 1649845482680, ssl.enabled.protocols = [TLSv1.2] 1649845482680, ssl.endpoint.identification.algorithm = null 1649845482680, ssl.key.password = null 1649845482680, ssl.keymanager.algorithm = SunX509 1649845482680, ssl.keystore.location = null 1649845482680, ssl.keystore.password = null 1649845482680, ssl.keystore.type = JKS 1649845482680, ssl.protocol = TLSv1.2 1649845482680, ssl.provider = null 1649845482680, ssl.secure.random.implementation = null 1649845482680, ssl.trustmanager.algorithm = PKIX 1649845482680, ssl.truststore.location = null 1649845482680, ssl.truststore.password = null 1649845482680, ssl.truststore.type = JKS 1649845482680, transaction.timeout.ms = 60000 1649845482680, transactional.id = null 1649845482680, value.serializer = class org.apache.kafka.common.serialization.StringSerializer 1649845482681,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-257] Instantiated an idempotent producer. 1649845482682,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-257] Overriding the default retries config to the recommended value of 2147483647 since the idempotent producer is enabled. 1649845482682,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-257] Overriding the default acks to all since idempotence is enabled. 1649845482682,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka version: 2.5.0 1649845482682,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka commitId: 66563e712b0b9f84 1649845482682,10:24:42 [] [] [] [] INFO ppInfoParser$AppInfo Kafka startTimeMs: 1649845482682 1649845482684,10:24:42 [] [] [] [] INFO Metadata [Producer clientId=producer-257] Cluster ID: G1_s3A4NTtuR07X2sF0UCg 1649845482684,10:24:42 [] [] [] [] INFO TransactionManager [Producer clientId=producer-257] ProducerId set to 148844 with epoch 0 1649845482689,10:24:42 [] [] [] [] ERROR InstanceEventHandler Error updating inventory Instance 1649845482689,java.lang.NumberFormatException: null 1649845482689, at java.lang.Integer.parseInt(Integer.java:614) ~[?:?] 1649845482689, at java.lang.Integer.parseInt(Integer.java:770) ~[?:?] 1649845482689, at org.folio.inventory.support.InstanceUtil.mergeFieldsWhichAreNotControlled(InstanceUtil.java:44) ~[ms.jar:?] 1649845482689, at org.folio.inventory.dataimport.handlers.actions.ReplaceInstanceEventHandler.prepareAndExecuteMapping(ReplaceInstanceEventHandler.java:170) ~[ms.jar:?] 1649845482689, at org.folio.inventory.dataimport.handlers.actions.ReplaceInstanceEventHandler.lambda$handle$0(ReplaceInstanceEventHandler.java:83) ~[ms.jar:?] 1649845482689, at java.util.Optional.map(Optional.java:265) ~[?:?] 1649845482689, at org.folio.inventory.dataimport.handlers.actions.ReplaceInstanceEventHandler.lambda$handle$2(ReplaceInstanceEventHandler.java:83) ~[ms.jar:?] 1649845482689, at io.vertx.core.impl.future.Composition.onSuccess(Composition.java:38) ~[ms.jar:?] 1649845482689, at io.vertx.core.impl.future.FutureBase.emitSuccess(FutureBase.java:62) ~[ms.jar:?] 1649845482689, at io.vertx.core.impl.future.FutureImpl.tryComplete(FutureImpl.java:179) ~[ms.jar:?] 1649845482689, at io.vertx.core.impl.future.PromiseImpl.tryComplete(PromiseImpl.java:23) ~[ms.jar:?] 1649845482689, at io.vertx.core.Promise.complete(Promise.java:66) ~[ms.jar:?] 1649845482689, at io.vertx.core.Future.lambda$fromCompletionStage$3(Future.java:384) ~[ms.jar:?] 1649845482689, at java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859) ~[?:?] 1649845482689, at java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837) ~[?:?] 1649845482689, at java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) ~[?:?] 1649845482689, at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:2073) ~[?:?] 1649845482689, at io.vertx.ext.web.client.impl.HttpContext.handleDispatchResponse(HttpContext.java:369) ~[ms.jar:?] 1649845482689, at io.vertx.ext.web.client.impl.HttpContext.execute(HttpContext.java:356) ~[ms.jar:?] 1649845482689, at io.vertx.ext.web.client.impl.HttpContext.next(HttpContext.java:334) ~[ms.jar:?] 1649845482689, at io.vertx.ext.web.client.impl.HttpContext.fire(HttpContext.java:301) ~[ms.jar:?] 1649845482689, at io.vertx.ext.web.client.impl.HttpContext.dispatchResponse(HttpContext.java:263) ~[ms.jar:?] 1649845482689, at io.vertx.ext.web.client.impl.HttpContext.lambda$null$8(HttpContext.java:533) ~[ms.jar:?] 1649845482689, at io.vertx.core.impl.AbstractContext.dispatch(AbstractContext.java:96) ~[ms.jar:?] 1649845482689, at io.vertx.core.impl.WorkerContext.lambda$run$1(WorkerContext.java:102) ~[ms.jar:?] 1649845482689, at io.vertx.core.impl.TaskQueue.run(TaskQueue.java:76) ~[ms.jar:?] 1649845482689, at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) [?:?] 1649845482689, at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) [?:?] 1649845482689, at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) [ms.jar:?] 1649845482689, at java.lang.Thread.run(Thread.java:829) [?:?] 1649845482736,10:24:42 [] [] [] [] INFO KafkaEventPublisher Event with type: DI_ERROR and recordId: 6b698831-3ed8-4c66-87dc-5b5e21b4677a was sent to the topic kmsd.Default.fs00001019.DI_ERROR 1649845482736,"10:24:42 [] [] [] [] ERROR KafkaConsumerWrapper Error while processing a record - id: 32 subscriptionPattern: SubscriptionDefinition(eventType=DI_INVENTORY_INSTANCE_MATCHED, subscriptionPattern=kmsd\.Default\.\w{1,}\.DI_INVENTORY_INSTANCE_MATCHED)" 1649845482736,io.vertx.core.impl.NoStackTraceThrowable: Failed to process data import event payload 1649845482744,10:24:42 [] [] [] [] INFO KafkaEventPublisher Event with type: DI_ERROR and recordId: 96ef2306-67c7-4c53-bd48-6e73b7612686 was sent to the topic kmsd.Default.fs00001019.DI_ERROR 1649845482744,"10:24:42 [] [] [] [] ERROR KafkaConsumerWrapper Error while processing a record - id: 32 subscriptionPattern: SubscriptionDefinition(eventType=DI_INVENTORY_INSTANCE_MATCHED, subscriptionPattern=kmsd\.Default\.\w{1,}\.DI_INVENTORY_INSTANCE_MATCHED)" 1649845482744,io.vertx.core.impl.NoStackTraceThrowable: Failed to process data import event payload 1649845482744,10:24:42 [] [] [] [] INFO KafkaProducer [Producer clientId=producer-257] Closing the Kafka producer with timeoutMillis = 9223372036854775807 ms. 1649845483230,"10:24:43 [] [] [] [] INFO KafkaConsumerWrapper Consumer - id: 8 subscriptionPattern: SubscriptionDefinition(eventType=DI_SRS_MARC_BIB_RECORD_CREATED, subscriptionPattern=kmsd\.Default\.\w{1,}\.DI_SRS_MARC_BIB_RECORD_CREATED) Committed offset: 286" 1649845483235,"10:24:43 [] [] [] [] INFO KafkaConsumerWrapper Consumer - id: 8 subscriptionPattern: SubscriptionDefinition(eventType=DI_SRS_MARC_BIB_RECORD_CREATED, subscriptionPattern=kmsd\.Default\.\w{1,}\.DI_SRS_MARC_BIB_RECORD_CREATED) Committed offset: 287" 1649845483674,"10:24:43 [] [] [] [] INFO KafkaConsumerWrapper Consumer - id: 32 subscriptionPattern: SubscriptionDefinition(eventType=DI_INVENTORY_INSTANCE_MATCHED, subscriptionPattern=kmsd\.Default\.\w{1,}\.DI_INVENTORY_INSTANCE_MATCHED) Committed offset: 54" 1649845483678,"10:24:43 [] [] [] [] INFO KafkaConsumerWrapper Consumer - id: 32 subscriptionPattern: SubscriptionDefinition(eventType=DI_INVENTORY_INSTANCE_MATCHED, subscriptionPattern=kmsd\.Default\.\w{1,}\.DI_INVENTORY_INSTANCE_MATCHED) Committed offset: 55"