|
13
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.Dataset.isEmpty(Dataset.scala:653)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.readInternalStatisticalForecast(DemandDetermination.java:114)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:98)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:54
|
0 ms
|
|
|
|
14
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.Dataset.isEmpty(Dataset.scala:653)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.readExternalAOF(DemandDetermination.java:122)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:99)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:54
|
0 ms
|
|
|
|
15
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.Dataset.isEmpty(Dataset.scala:653)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.readExternalStatisticalForecast(DemandDetermination.java:131)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:100)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:54
|
1 ms
|
|
|
|
16
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.Dataset.isEmpty(Dataset.scala:653)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.readCustomerOrders(DemandDetermination.java:139)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:101)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:54
|
0 ms
|
|
|
|
17
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.Dataset.isEmpty(Dataset.scala:653)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.insights.InsightCreatorImpl.createNoDataAvailableInsight(InsightCreatorImpl.java:38)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.storeBufferedInsights(DemandDetermination.java:149)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:103)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:54
|
0 ms
|
|
|
|
18
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:243)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.FileStorageHandler.writeToStorage(FileStorageHandler.java:96)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.insights.BufferedInsightManager.writeInsightsToStorage(BufferedInsightManager.java:68)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.storeBufferedInsights(DemandDetermination.java:150)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:103)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2025/12/18 04:04:54
|
3 s
|
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
19
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:243)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.FileStorageHandler.writeToStorage(FileStorageHandler.java:96)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.insights.BufferedInsightManager.writeInsightsToStorage(BufferedInsightManager.java:68)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.storeBufferedInsights(DemandDetermination.java:150)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:103)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2025/12/18 04:04:55
|
0.6 s
|
[24]
|
|
20
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:1006)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:1005)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:165)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles(OptimisticTransaction.scala:1002)
org.apache.spark.sql.delta.OptimisticTransactionImpl.filterFiles$(OptimisticTransaction.scala:1002)
org.apache.spark.sql.delta.OptimisticTransaction.filterFiles(OptimisticTransaction.scala:165)
org.apache.spark.sql.delta.commands.WriteIntoDelta.writeAndReturnCommitData(WriteIntoDelta.scala:335)
org.apache.spark.sql.delta.commands.WriteIntoDelta.$anonfun$run$1(WriteIntoDelta.scala:108)
org.apache.spark.sql.delta.DeltaLog.withNewTransaction(DeltaLog.scala:241)
|
2025/12/18 04:04:55
|
31 ms
|
[25]
|
|
21
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.OptimisticTransaction.recordOperation(OptimisticTransaction.scala:165)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.OptimisticTransaction.recordDeltaOperation(OptimisticTransaction.scala:165)
org.apache.spark.sql.delta.OptimisticTransactionImpl.$anonfun$doCommitRetryIteratively$1(OptimisticTransaction.scala:2223)
org.apache.spark.sql.delta.SnapshotManagement.withSnapshotLockInterruptibly(SnapshotManagement.scala:85)
org.apache.spark.sql.delta.SnapshotManagement.withSnapshotLockInterruptibly$(SnapshotManagement.scala:82)
org.apache.spark.sql.delta.DeltaLog.withSnapshotLockInterruptibly(DeltaLog.scala:75)
org.apache.spark.sql.delta.OptimisticTransactionImpl.lockCommitIfEnabled(OptimisticTransaction.scala:2192)
org.apache.spark.sql.delta.OptimisticTransactionImpl.doCommitRetryIteratively(OptimisticTransaction.scala:2210)
org.apache.spark.sql.delta.OptimisticTransactionImpl.doCommitRetryIteratively$(OptimisticTransaction.scala:2206)
org.apache.spark.sql.delta.OptimisticTransaction.doCommitRetryIteratively(OptimisticTransaction.scala:165)
org.apache.spark.sql.delta.OptimisticTransactionImpl.liftedTree1$1(OptimisticTransaction.scala:1434)
org.apache.spark.sql.delta.OptimisticTransactionImpl.$anonfun$commitImpl$1(OptimisticTransaction.scala:1330)
org.apache.spark.sql.delta.metering.DeltaLogging.recordFrameProfile(DeltaLogging.scala:171)
org.apache.spark.sql.delta.metering.DeltaLogging.recordFrameProfile$(DeltaLogging.scala:169)
|
2025/12/18 04:04:56
|
86 ms
|
[26]
|
|
|
0
|
replenishmentRunId = 10000003975 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 8eadde78-f596-3022-9328-ac02349dc091 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 067fd3f9-4357-3330-a6b6-aa74df43090f attempt = 1 cornerstoneTenantId = 7354 marketUnit = Daily_Grocery scenario = STANDARD
+details
replenishmentRunId = 10000003975 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 8eadde78-f596-3022-9328-ac02349dc091 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 067fd3f9-4357-3330-a6b6-aa74df43090f attempt = 1 cornerstoneTenantId = 7354 marketUnit = Daily_Grocery scenario = STANDARD org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.KafkaMessagePublisher.publishMessage(KafkaMessagePublisher.java:116)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.KafkaMessagePublisher.publishInsight(KafkaMessagePublisher.java:49)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.insights.InsightHandler.processAndPublishInsights(InsightHandler.java:17)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.NotifyDemandServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyDemandServiceReplRunFinishedActivityImpl.java:35)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/12 09:10:28
|
0.2 s
|
[0]
|
|
|
1
|
replenishmentRunId = 10000003976 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = d281388d-eb56-314b-aa27-ce61fb479456 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = c865f363-6bf7-3901-9a0c-bb29b6e15db7 attempt = 1 cornerstoneTenantId = 7354 marketUnit = Daily_Grocery scenario = STANDARD
+details
replenishmentRunId = 10000003976 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = d281388d-eb56-314b-aa27-ce61fb479456 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = c865f363-6bf7-3901-9a0c-bb29b6e15db7 attempt = 1 cornerstoneTenantId = 7354 marketUnit = Daily_Grocery scenario = STANDARD org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.KafkaMessagePublisher.publishMessage(KafkaMessagePublisher.java:116)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.KafkaMessagePublisher.publishInsight(KafkaMessagePublisher.java:49)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.insights.InsightHandler.processAndPublishInsights(InsightHandler.java:17)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.NotifyDemandServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyDemandServiceReplRunFinishedActivityImpl.java:35)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/12 09:14:54
|
17 ms
|
[1]
|
|
|
2
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.Dataset.isEmpty(Dataset.scala:653)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandProfileDetermination.determineDemandProfileData(DemandProfileDetermination.java:34)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.WorkloadDetermination.determineWorkload(WorkloadDetermination.java:66)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:95)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:30
|
5 s
|
[4][5][6]
|
|
|
3
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.delta.util.threads.DeltaThreadPool.$anonfun$submit$1(DeltaThreadPool.scala:39)
java.base/java.util.concurrent.FutureTask.run(Unknown Source)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.$anonfun$run$1(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:77)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured$(SparkThreadLocalForwardingThreadPoolExecutor.scala:60)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:116)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.run(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:37
|
0.4 s
|
[7]
|
|
|
4
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:677)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:677)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:922)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:942)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4860)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3551)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2302)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2177)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2068)
com.google.common.cache.LocalCache.get(LocalCache.java:3986)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4855)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:941)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:951)
org.apache.spark.sql.delta.DeltaLog$.forTable(DeltaLog.scala:782)
|
2025/12/18 04:04:38
|
0.7 s
|
[8]
|
|
|
5
|
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 31
+details
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 31 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:04:39
|
2 s
|
[9]
|
|
|
6
|
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 31
+details
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 31 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:04:42
|
3 s
|
[10][11]
|
|
|
7
|
Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:04:45
|
0.7 s
|
[12]
|
|
|
8
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:677)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:677)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:922)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:942)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4860)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3551)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2302)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2177)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2068)
com.google.common.cache.LocalCache.get(LocalCache.java:3986)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4855)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:941)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:951)
org.apache.spark.sql.delta.DeltaLog$.forTable(DeltaLog.scala:782)
|
2025/12/18 04:04:46
|
0.3 s
|
[13]
|
|
|
9
|
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1
+details
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:04:47
|
0.5 s
|
[14]
|
|
|
10
|
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1
+details
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:04:48
|
2 s
|
[15][16]
|
|
|
11
|
Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:04:49
|
0.4 s
|
[17]
|
|
|
12
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.Dataset.isEmpty(Dataset.scala:653)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandProfileDetermination.determineDemandProfileData(DemandProfileDetermination.java:39)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.WorkloadDetermination.determineWorkload(WorkloadDetermination.java:66)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:95)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:50
|
4 s
|
[18][19][20][21][22][23]
|
|
|
22
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:243)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.FileStorageHandler.writeToStorage(FileStorageHandler.java:96)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.FileStorageHandler.writeToStorage(FileStorageHandler.java:81)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.DemandResultWriterImpl.writeDemandToStorage(DemandResultWriterImpl.java:36)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.DemandDetermination.determineDemands(DemandDetermination.java:105)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:57)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2025/12/18 04:04:57
|
2 s
|
[27][28][29]
|
|
|
23
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = GetDemands activityId = 80ab2e30-7ec4-39fd-8c9f-3de75f2c4bf4 workflowType = GetDemandsWorkflow workflowId = 5731084e-62a2-34e7-abfe-3ab62c0765d1 attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:243)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.FileStorageHandler.writePerformanceMetricsToStorage(FileStorageHandler.java:305)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.SparkMetricsWriterImpl.storePerformanceMetrics(SparkMetricsWriterImpl.java:17)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.GetDemandsActivityImpl.getDemands(GetDemandsActivityImpl.java:58)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:04:59
|
2 s
|
[30][31][32][33][34]
|
|
|
24
|
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 092ccc0e-e64f-328a-9cf3-e1b42af4bfb5 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 6f7617c2-d45a-3c83-8d7f-00e8e16dd2de attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD
+details
replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 092ccc0e-e64f-328a-9cf3-e1b42af4bfb5 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 6f7617c2-d45a-3c83-8d7f-00e8e16dd2de attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:251)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.KafkaMessagePublisher.publishMessage(KafkaMessagePublisher.java:116)
com.sap.s4hana.eureka.business.crpdemandservice.storageaccess.KafkaMessagePublisher.publishInsight(KafkaMessagePublisher.java:49)
com.sap.s4hana.eureka.business.crpdemandservice.core.business.insights.InsightHandler.processAndPublishInsights(InsightHandler.java:17)
com.sap.s4hana.eureka.business.crpdemandservice.core.controller.NotifyDemandServiceReplRunFinishedActivityImpl.notifyReplRunFinished(NotifyDemandServiceReplRunFinishedActivityImpl.java:35)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2025/12/18 04:08:38
|
2 s
|
[39]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
25
|
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 092ccc0e-e64f-328a-9cf3-e1b42af4bfb5 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 6f7617c2-d45a-3c83-8d7f-00e8e16dd2de attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 0
+details
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 092ccc0e-e64f-328a-9cf3-e1b42af4bfb5 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 6f7617c2-d45a-3c83-8d7f-00e8e16dd2de attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 0 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:08:38
|
0.3 s
|
[35]
|
|
26
|
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 092ccc0e-e64f-328a-9cf3-e1b42af4bfb5 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 6f7617c2-d45a-3c83-8d7f-00e8e16dd2de attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 0
+details
Delta: Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 092ccc0e-e64f-328a-9cf3-e1b42af4bfb5 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 6f7617c2-d45a-3c83-8d7f-00e8e16dd2de attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query: Compute snapshot for version: 0 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:08:38
|
1 s
|
[36][37]
|
|
27
|
Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 092ccc0e-e64f-328a-9cf3-e1b42af4bfb5 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 6f7617c2-d45a-3c83-8d7f-00e8e16dd2de attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000003977 tenantId = 8322290869003019352 activityType = NotifyReplRunFinished activityId = 092ccc0e-e64f-328a-9cf3-e1b42af4bfb5 workflowType = NotifyDemandServiceReplRunFinishedWorkflow workflowId = 6f7617c2-d45a-3c83-8d7f-00e8e16dd2de attempt = 1 cornerstoneTenantId = 7354 marketUnit = NML_DLY scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1187)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1184)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:81)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:132)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:117)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:112)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:306)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:150)
|
2025/12/18 04:08:39
|
0.3 s
|
[38]
|
|