Expand description
Defines all kinds of node in the plan tree, each node represent a relational expression.
We use a immutable style tree structure, every Node are immutable and cannot be modified after
it has been created. If you want to modify the node, such as rewriting the expression in a
ProjectNode
or changing a node’s input node, you need to create a new node. We use Rc as the
node’s reference, and a node just storage its inputs’ reference, so change a node just need
create one new node but not the entire sub-tree.
So when you want to add a new node, make sure:
- each field in the node struct are private
- recommend to implement the construction of Node in a unified
new()
function, if have multi methods to construct, make they have a consistent behavior - all field should be valued in construction, so the properties’ derivation should be finished
in the
new()
function.
Re-exports§
pub use generic::PlanAggCall;
pub use generic::PlanAggCallDisplay;
Modules§
- batch
- batch_
delete 🔒 - batch_
exchange 🔒 - batch_
expand 🔒 - batch_
file_ 🔒scan - batch_
filter 🔒 - batch_
group_ 🔒topn - batch_
hash_ 🔒agg - batch_
hash_ 🔒join - batch_
hop_ 🔒window - batch_
iceberg_ 🔒scan - batch_
insert 🔒 - batch_
kafka_ 🔒scan - batch_
limit 🔒 - batch_
log_ 🔒seq_ scan - batch_
lookup_ 🔒join - batch_
max_ 🔒one_ row - batch_
mysql_ 🔒query - batch_
nested_ 🔒loop_ join - batch_
over_ 🔒window - batch_
postgres_ 🔒query - batch_
project 🔒 - batch_
project_ 🔒set - batch_
seq_ 🔒scan - batch_
simple_ 🔒agg - batch_
sort 🔒 - batch_
sort_ 🔒agg - batch_
source 🔒 - batch_
sys_ 🔒seq_ scan - batch_
table_ 🔒function - batch_
topn 🔒 - batch_
union 🔒 - batch_
update 🔒 - batch_
values 🔒 - col_
pruning 🔒 - convert 🔒
- derive 🔒
- eq_
join_ 🔒predicate - expr_
rewritable 🔒 - expr_
visitable 🔒 - generic
- This module contains the generic plan nodes that are shared by all the plan nodes. They are meant to reuse the common fields between logical, batch and stream nodes.
- logical_
agg 🔒 - logical_
apply 🔒 - logical_
cdc_ 🔒scan - logical_
changelog 🔒 - logical_
cte_ 🔒ref - logical_
dedup 🔒 - logical_
delete 🔒 - logical_
except 🔒 - logical_
expand 🔒 - logical_
file_ 🔒scan - logical_
filter 🔒 - logical_
hop_ 🔒window - logical_
iceberg_ 🔒scan - logical_
insert 🔒 - logical_
intersect 🔒 - logical_
join 🔒 - logical_
kafka_ 🔒scan - logical_
limit 🔒 - logical_
max_ 🔒one_ row - logical_
multi_ 🔒join - logical_
mysql_ 🔒query - logical_
now 🔒 - logical_
over_ 🔒window - logical_
postgres_ 🔒query - logical_
project 🔒 - logical_
project_ 🔒set - logical_
recursive_ 🔒union - logical_
scan 🔒 - logical_
share 🔒 - logical_
source 🔒 - logical_
sys_ 🔒scan - logical_
table_ 🔒function - logical_
topn 🔒 - logical_
union 🔒 - logical_
update 🔒 - logical_
values 🔒 - merge_
eq_ 🔒nodes - plan_
base 🔒 - plan_
node_ 🔒meta - plan_
tree_ 🔒node - predicate_
pushdown 🔒 - stream
- stream_
asof_ 🔒join - stream_
cdc_ 🔒table_ scan - stream_
changelog 🔒 - stream_
dedup 🔒 - stream_
delta_ 🔒join - stream_
dml 🔒 - stream_
dynamic_ 🔒filter - stream_
eowc_ 🔒over_ window - stream_
exchange 🔒 - stream_
expand 🔒 - stream_
filter 🔒 - stream_
fs_ 🔒fetch - stream_
global_ 🔒approx_ percentile - stream_
group_ 🔒topn - stream_
hash_ 🔒agg - stream_
hash_ 🔒join - stream_
hop_ 🔒window - stream_
join_ 🔒common - stream_
local_ 🔒approx_ percentile - stream_
materialize 🔒 - stream_
now 🔒 - stream_
over_ 🔒window - stream_
project 🔒 - stream_
project_ 🔒set - stream_
row_ 🔒id_ gen - stream_
row_ 🔒merge - stream_
share 🔒 - stream_
simple_ 🔒agg - stream_
sink 🔒 - stream_
sort 🔒 - stream_
source 🔒 - stream_
source_ 🔒scan - stream_
stateless_ 🔒simple_ agg - stream_
sync_ 🔒log_ store - stream_
table_ 🔒scan - stream_
temporal_ 🔒join - stream_
topn 🔒 - stream_
union 🔒 - stream_
values 🔒 - stream_
watermark_ 🔒filter - to_
prost 🔒 - utils
Macros§
- impl_
down_ 🔒cast_ fn - impl plan node downcast fn for each node.
- impl_
plan_ 🔒node - impl_
plan_ 🔒node_ meta - impl
PlanNodeType
fn for each node.
Structs§
- Batch
- The marker for batch convention.
- Batch
Delete BatchDelete
implementssuper::LogicalDelete
- Batch
Exchange BatchExchange
imposes a particular distribution on its input without changing its content.- Batch
Expand - Batch
Extra - Extra fields for batch plan nodes.
- Batch
File Scan - Batch
Filter BatchFilter
implementssuper::LogicalFilter
- Batch
Group TopN BatchGroupTopN
implementssuper::LogicalTopN
to find the top N elements with a heap- Batch
Hash Agg - Batch
Hash Join BatchHashJoin
implementssuper::LogicalJoin
with hash table. It builds a hash table from inner (right-side) relation and then probes with data from outer (left-side) relation to get output rows.- Batch
HopWindow BatchHopWindow
implementssuper::LogicalHopWindow
to evaluate specified expressions on input rows- Batch
Iceberg Scan - Batch
Insert BatchInsert
implementssuper::LogicalInsert
- Batch
Kafka Scan - Batch
Limit BatchLimit
implementssuper::LogicalLimit
to fetch specified rows from input- Batch
LogSeq Scan - Batch
Lookup Join - Batch
MaxOne Row BatchMaxOneRow
fetches up to one row from the input, returning an error if the input contains more than one row at runtime.- Batch
MySql Query - Batch
Nested Loop Join BatchNestedLoopJoin
implementssuper::LogicalJoin
by checking the join condition against all pairs of rows from inner & outer side within 2 layers of loops.- Batch
Over Window - Batch
Postgres Query - Batch
Project BatchProject
implementssuper::LogicalProject
to evaluate specified expressions on input rows- Batch
Project Set - Batch
SeqScan BatchSeqScan
implementssuper::LogicalScan
to scan from a row-oriented table- Batch
Simple Agg - Batch
Sort BatchSort
buffers all data from input and sort these rows by specified order, providing the collation required by user or parent plan node.- Batch
Sort Agg - Batch
Source BatchSource
represents a table/connector source at the very beginning of the graph.- Batch
SysSeq Scan BatchSysSeqScan
implementssuper::LogicalSysScan
to scan from a row-oriented table- Batch
Table Function - Batch
TopN BatchTopN
implementssuper::LogicalTopN
to find the top N elements with a heap- Batch
Union BatchUnion
implementssuper::LogicalUnion
- Batch
Update BatchUpdate
implementssuper::LogicalUpdate
- Batch
Values - Column
Pruning Context - EqJoin
Predicate - The join predicate used in optimizer
- EqJoin
Predicate Display - Iceberg
Partition Info - Logical
- The marker for logical convention.
- Logical
Agg LogicalAgg
groups input data by their group key and computes aggregation functions.- Logical
Apply LogicalApply
represents a correlated join, where the right side may refer to columns from the left side.- Logical
CdcScan LogicalCdcScan
reads rows of a table from an external upstream database- Logical
Change Log - Logical
CteRef - Logical
Dedup LogicalDedup
deduplicates data on specific columns. It is now used inDISTINCT ON
without anORDER BY
.- Logical
Delete LogicalDelete
iterates on input relation and delete the data from specified table.- Logical
Except LogicalExcept
returns the rows of its first input except any matching rows from its other inputs.- Logical
Expand LogicalExpand
expands one row multiple times according tocolumn_subsets
and also keeps original columns of input. It can be used to implement distinct aggregation and group set.- Logical
File Scan - Logical
Filter LogicalFilter
iterates over its input and returns elements for whichpredicate
evaluates to true, filtering out the others.- Logical
HopWindow LogicalHopWindow
implements Hop Table Function.- Logical
Iceberg Scan LogicalIcebergScan
is only used by batch queries. At the beginning of the batch query optimization,LogicalSource
with a iceberg property would be converted into aLogicalIcebergScan
.- Logical
Insert LogicalInsert
iterates on input relation and insert the data into specified table.- Logical
Intersect LogicalIntersect
returns the intersect of the rows of its inputs. Ifall
is false, it needs to eliminate duplicates.- Logical
Join LogicalJoin
combines two relations according to some condition.- Logical
Kafka Scan LogicalKafkaScan
is only used by batch queries. At the beginning of the batch query optimization,LogicalSource
with a kafka property would be converted into aLogicalKafkaScan
.- Logical
Limit LogicalLimit
fetches up tolimit
rows fromoffset
- Logical
MaxOne Row LogicalMaxOneRow
fetches up to one row from the input, returning an error if the input contains more than one row at runtime. Only available in batch mode.- Logical
Multi Join LogicalMultiJoin
combines two or more relations according to some condition.- Logical
Multi Join Builder - Logical
MySql Query - Logical
Now - Logical
Over Window LogicalOverWindow
performsOVER
window functions to its input.- Logical
Postgres Query - Logical
Project LogicalProject
computes a set of expressions from its input relation.- Logical
Project Set LogicalProjectSet
projects one row multiple times according toselect_list
.- Logical
Recursive Union LogicalRecursiveUnion
returns the union of the rows of its inputs. note: ifall
is false, it needs to eliminate duplicates.- Logical
Scan LogicalScan
returns contents of a table or other equivalent object- Logical
Share LogicalShare
operator is used to represent reusing of existing operators. It is the key operator for DAG plan. It could have multiple parents which makes it different from other operators. Currently, it has been used to the following scenarios:- Logical
Source LogicalSource
returns contents of a table or other equivalent object- Logical
SysScan LogicalSysScan
returns contents of a table or other equivalent object- Logical
Table Function LogicalTableFunction
is a scalar/table function used as a relation (in theFROM
clause).- Logical
TopN LogicalTopN
sorts the input data and fetches up tolimit
rows fromoffset
- Logical
Union LogicalUnion
returns the union of the rows of its inputs. Ifall
is false, it needs to eliminate duplicates.- Logical
Update LogicalUpdate
iterates on input relation, set some columns, and inject update records into specified table.- Logical
Values LogicalValues
builds rows according to a list of expressions- NoExtra
- No extra fields for logical plan nodes.
- Plan
Base - The common fields of all plan nodes with different conventions.
- Plan
Node Id - PlanRef
- Predicate
Pushdown Context - Rewrite
Stream Context - Stream
- The marker for stream convention.
- Stream
AsOf Join StreamAsOfJoin
implementssuper::LogicalJoin
with hash tables.- Stream
CdcTable Scan StreamCdcTableScan
is a virtual plan node to represent a stream cdc table scan. It will be converted to cdc backfill + merge node (for upstream source)- Stream
Change Log - Stream
Dedup - Stream
Delta Join StreamDeltaJoin
implementssuper::LogicalJoin
with delta join. It requires its two inputs to be indexes.- Stream
Dml - Stream
Dynamic Filter - Stream
Eowc Over Window - Stream
Eowc Sort - Stream
Exchange StreamExchange
imposes a particular distribution on its input without changing its content.- Stream
Expand - Stream
Extra - Extra fields for stream plan nodes.
- Stream
Filter StreamFilter
implementssuper::LogicalFilter
- Stream
FsFetch - Fetch files from filesystem/s3/iceberg.
- Stream
Global Approx Percentile - Stream
Group TopN - Stream
Hash Agg - Stream
Hash Join StreamHashJoin
implementssuper::LogicalJoin
with hash table. It builds a hash table from inner (right-side) relation and probes with data from outer (left-side) relation to get output rows.- Stream
HopWindow StreamHopWindow
represents a hop window table function.- Stream
Local Approx Percentile - Stream
Materialize - Materializes a stream.
- Stream
Now - Stream
Over Window - Stream
Project StreamProject
implementssuper::LogicalProject
to evaluate specified expressions on input rows.- Stream
Project Set - Stream
RowId Gen - Stream
RowMerge StreamRowMerge
is used for merging two streams with the same stream key and distribution. It will buffer the outputs from its input streams until we receive a barrier. On receiving a barrier, it willProject
their outputs according to the providedlhs_mapping
andrhs_mapping
.- Stream
Share StreamShare
will be translated into anExchangeNode
based on its distribution finally.- Stream
Simple Agg - Stream
Sink StreamSink
represents a table/connector sink at the very end of the graph.- Stream
Source StreamSource
represents a table/connector source at the very beginning of the graph.- Stream
Source Scan StreamSourceScan
scans from a shared source. It forwards data from the upstreamStreamSource
, and also backfills data from the external source.- Stream
Stateless Simple Agg - Streaming stateless simple agg.
- Stream
Sync LogStore - Stream
Table Scan StreamTableScan
is a virtual plan node to represent a stream table scan. It will be converted to stream scan + merge node (for upstream materialize) + batch table scan when converting toMView
creation request.- Stream
Temporal Join - Stream
TopN StreamTopN
implementssuper::LogicalTopN
to find the top N elements with a heap- Stream
Union StreamUnion
implementssuper::LogicalUnion
- Stream
Values StreamValues
implementsLogicalValues.to_stream()
- Stream
Watermark Filter - ToStream
Context
Enums§
- Convention
- Partition
Compute Info - Why we need
PartitionComputeInfo
? - Plan
Base Ref - Reference to
PlanBase
with erased conventions. - Plan
Node Type - each enum value represent a PlanNode struct type, help us to dispatch and downcast
Constants§
- PLAN_
DEPTH_ THRESHOLD - Recursion depth threshold for plan node visitor to send notice to user.
- PLAN_
TOO_ DEEP_ NOTICE - Notice message for plan node visitor to send to user when the depth threshold is reached.
Traits§
- ColPrunable
- The trait for column pruning, only logical plan node will use it, though all plan node impl it.
- Convention
Marker - A marker trait for different conventions, used for enforcing type safety.
- Endo
Plan - A more sophisticated
Endo
taking into account of the DAG structure ofPlanRef
. In addition toEndo
, one have to specify thecached
function to persist transformedLogicalShare
and their results, and thedag_apply
function will take care to only transform everyLogicalShare
nodes once. - Explain
- Expr
Rewritable - Rewrites expressions in a
PlanRef
. Due toShare
operator, theExprRewriter
needs to be idempotent i.e., applying it more than once to the sameExprImpl
will be a noop on subsequent applications.rewrite_exprs
should only return a plan with the given node modified. To rewrite recursively, callrewrite_exprs_recursive
onRewriteExprsRecursive
. - Plan
Node - The common trait over all plan nodes. Used by optimizer framework which will treat all node as
dyn PlanNode
- Plan
Node Meta - The trait for accessing the meta data and
PlanBase
for plan nodes. - Plan
Tree Node - The trait
PlanNode
really need about tree structure and used by optimizer framework. every plan node should impl it. - Plan
Tree Node Binary - See
PlanTreeNode
- Plan
Tree Node Leaf - See
PlanTreeNode
- Plan
Tree Node Unary - See
PlanTreeNode
- Predicate
Pushdown - The trait for predicate pushdown, only logical plan node will use it, though all plan node impl it.
- Rewrite
Exprs 🔒Recursive - Semantics
- Stream
Node - ToBatch
ToBatch
allows to convert a logical plan node to batch physical node with an optional required order.- ToBatch
Pb - ToDistributed
Batch ToDistributedBatch
allows to convert a batch physical plan to distributed batch plan, by insert exchange node, with an optional required order and distributed.- ToLocal
Batch - Converts a batch physical plan to local plan for local execution.
- ToPb
- ToStream
ToStream
converts a logical plan node to streaming physical node with an optional required distribution.- TryTo
Batch Pb - TryTo
Stream Pb - Visit
Exprs 🔒Recursive - Visit
Plan - A more sophisticated
Visit
taking into account of the DAG structure ofPlanRef
. In addition toVisit
, one have to specifyvisited
to store and report visitedLogicalShare
nodes, and thedag_visit
function will take care to only visit everyLogicalShare
nodes once. See alsoEndoPlan
.
Functions§
- gen_
filter_ and_ pushdown - pretty_
config 🔒 - reorganize_
elements_ id - In order to let expression display id started from 1 for explaining, hidden column names and other places. We will reset expression display id to 0 and clone the whole plan to reset the schema.
- stream_
enforce_ eowc_ requirement