digraph G {
0 [labelType="html" label="<br><b>Union</b><br><br>"];
subgraph cluster1 {
isCluster="true";
label="WholeStageCodegen (1)\n \nduration: 301 ms";
2 [labelType="html" label="<br><b>Project</b><br><br>"];
3 [labelType="html" label="<b>Filter</b><br><br>number of output rows: 2"];
4 [labelType="html" label="<b>ColumnarToRow</b><br><br>number of output rows: 12<br>number of input batches: 1"];
}
5 [labelType="html" label="<b>Scan parquet </b><br><br>number of files read: 1<br>scan time: 300 ms<br>dynamic partition pruning time: 0 ms<br>metadata time: 15 ms<br>size of files read: 19.4 KiB<br>number of output rows: 12<br>number of partitions read: 1"];
subgraph cluster6 {
isCluster="true";
label="WholeStageCodegen (2)\n \nduration: 222 ms";
7 [labelType="html" label="<br><b>Project</b><br><br>"];
8 [labelType="html" label="<b>Filter</b><br><br>number of output rows: 0"];
}
9 [labelType="html" label="<b>Scan json </b><br><br>number of files read: 1<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 1913.0 B<br>number of output rows: 3<br>number of partitions read: 1"];
2->0;
3->2;
4->3;
5->4;
7->0;
8->7;
9->8;
}
10
Union
Project [protocol#193, metaData#192, commitInfo#194.inCommitTimestamp AS inCommitTimestamp#250L, 30 AS version#222L]
Filter (isnotnull(protocol#193.minReaderVersion) OR isnotnull(metaData#192.id))
ColumnarToRow
WholeStageCodegen (1)
FileScan parquet [metaData#192,protocol#193,commitInfo#194,version#195L] Batched: true, DataFilters: [(isnotnull(protocol#193.minReaderVersion) OR isnotnull(metaData#192.id))], Format: Parquet, Location: DeltaLogFileIndex(1 paths)[hdlfs://7646b954-15f6-4bdc-91a5-2644c1a43a19.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [Or(IsNotNull(protocol.minReaderVersion),IsNotNull(metaData.id))], ReadSchema: struct<metaData:struct<id:string,name:string,description:string,format:struct<provider:string,opt...
Project [protocol#232, metaData#231, commitInfo#233.inCommitTimestamp AS inCommitTimestamp#273L, version#234L]
Filter ((isnotnull(protocol#232.minReaderVersion) OR isnotnull(metaData#231.id)) OR (isnotnull(commitInfo#233.inCommitTimestamp) AND (version#234L = 31)))
WholeStageCodegen (2)
FileScan json [metaData#231,protocol#232,commitInfo#233,version#234L] Batched: false, DataFilters: [((isnotnull(protocol#232.minReaderVersion) OR isnotnull(metaData#231.id)) OR isnotnull(commitInf..., Format: JSON, Location: DeltaLogFileIndex(1 paths)[hdlfs://7646b954-15f6-4bdc-91a5-2644c1a43a19.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<metaData:struct<id:string,name:string,description:string,format:struct<provider:string,opt...
== Physical Plan ==
Union (8)
:- * Project (4)
: +- * Filter (3)
: +- * ColumnarToRow (2)
: +- Scan parquet (1)
+- * Project (7)
+- * Filter (6)
+- Scan json (5)
(1) Scan parquet
Output [4]: [metaData#192, protocol#193, commitInfo#194, version#195L]
Batched: true
Location: DeltaLogFileIndex [hdlfs://7646b954-15f6-4bdc-91a5-2644c1a43a19.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/prp/demand-profile-v3/_delta_log/00000000000000000030.checkpoint.parquet]
PushedFilters: [Or(IsNotNull(protocol.minReaderVersion),IsNotNull(metaData.id))]
ReadSchema: struct<metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,commitInfo:struct<inCommitTimestamp:bigint>>
(2) ColumnarToRow [codegen id : 1]
Input [4]: [metaData#192, protocol#193, commitInfo#194, version#195L]
(3) Filter [codegen id : 1]
Input [4]: [metaData#192, protocol#193, commitInfo#194, version#195L]
Condition : (isnotnull(protocol#193.minReaderVersion) OR isnotnull(metaData#192.id))
(4) Project [codegen id : 1]
Output [4]: [protocol#193, metaData#192, commitInfo#194.inCommitTimestamp AS inCommitTimestamp#250L, 30 AS version#222L]
Input [4]: [metaData#192, protocol#193, commitInfo#194, version#195L]
(5) Scan json
Output [4]: [metaData#231, protocol#232, commitInfo#233, version#234L]
Batched: false
Location: DeltaLogFileIndex [hdlfs://7646b954-15f6-4bdc-91a5-2644c1a43a19.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/prp/demand-profile-v3/_delta_log/00000000000000000031.json]
ReadSchema: struct<metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(6) Filter [codegen id : 2]
Input [4]: [metaData#231, protocol#232, commitInfo#233, version#234L]
Condition : ((isnotnull(protocol#232.minReaderVersion) OR isnotnull(metaData#231.id)) OR (isnotnull(commitInfo#233.inCommitTimestamp) AND (version#234L = 31)))
(7) Project [codegen id : 2]
Output [4]: [protocol#232, metaData#231, commitInfo#233.inCommitTimestamp AS inCommitTimestamp#273L, version#234L]
Input [4]: [metaData#231, protocol#232, commitInfo#233, version#234L]
(8) Union