| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.flink.table.planner.plan.nodes.logical |
| |
| import org.apache.flink.table.planner.calcite.FlinkTypeFactory.{isRowtimeIndicatorType, isTimestampLtzIndicatorType} |
| import org.apache.flink.table.planner.plan.nodes.FlinkConventions |
| import org.apache.flink.table.planner.plan.utils.MatchUtil |
| |
| import org.apache.calcite.plan._ |
| import org.apache.calcite.rel.`type`.RelDataType |
| import org.apache.calcite.rel.convert.ConverterRule |
| import org.apache.calcite.rel.core.Match |
| import org.apache.calcite.rel.logical.LogicalMatch |
| import org.apache.calcite.rel.{RelCollation, RelNode} |
| import org.apache.calcite.rex.RexNode |
| import org.apache.calcite.util.{ImmutableBitSet, Litmus} |
| |
| import java.util |
| |
| import scala.collection.JavaConversions._ |
| |
| class FlinkLogicalMatch( |
| cluster: RelOptCluster, |
| traitSet: RelTraitSet, |
| input: RelNode, |
| rowType: RelDataType, |
| pattern: RexNode, |
| strictStart: Boolean, |
| strictEnd: Boolean, |
| patternDefinitions: util.Map[String, RexNode], |
| measures: util.Map[String, RexNode], |
| after: RexNode, |
| subsets: util.Map[String, _ <: util.SortedSet[String]], |
| allRows: Boolean, |
| partitionKeys: ImmutableBitSet, |
| orderKeys: RelCollation, |
| interval: RexNode) |
| extends Match( |
| cluster, |
| traitSet, |
| input, |
| rowType, |
| pattern, |
| strictStart, |
| strictEnd, |
| patternDefinitions, |
| measures, |
| after, |
| subsets, |
| allRows, |
| partitionKeys, |
| orderKeys, |
| interval) |
| with FlinkLogicalRel { |
| |
| override def isValid(litmus: Litmus, context: RelNode.Context): Boolean = { |
| val inputContainsRowTimeLtz = input.getRowType.getFieldList.exists { field => |
| isRowtimeIndicatorType(field.getType) && isTimestampLtzIndicatorType(field.getType) |
| } |
| if (inputContainsRowTimeLtz) { |
| val containMatchRowTimeWithoutArgs = getMeasures.values().exists( |
| MatchUtil.isFinalOnMatchRowTimeWithoutArgs) |
| if (containMatchRowTimeWithoutArgs) { |
| litmus.fail( |
| "MATCH_ROWTIME(rowtimeField) should be used when input stream contains " + |
| "rowtime attribute with TIMESTAMP_LTZ type.\n" + |
| "Please pass rowtime attribute field as input argument of " + |
| "MATCH_ROWTIME(rowtimeField) function.") |
| } else { |
| litmus.succeed() |
| } |
| } else { |
| litmus.succeed() |
| } |
| } |
| |
| |
| override def copy(traitSet: RelTraitSet, inputs: util.List[RelNode]): RelNode = { |
| new FlinkLogicalMatch( |
| cluster, |
| traitSet, |
| inputs.get(0), |
| rowType, |
| pattern, |
| strictStart, |
| strictEnd, |
| getPatternDefinitions, |
| getMeasures, |
| getAfter, |
| getSubsets, |
| allRows, |
| getPartitionKeys, |
| getOrderKeys, |
| getInterval) |
| } |
| } |
| |
| private class FlinkLogicalMatchConverter |
| extends ConverterRule( |
| classOf[LogicalMatch], |
| Convention.NONE, |
| FlinkConventions.LOGICAL, |
| "FlinkLogicalMatchConverter") { |
| |
| override def convert(rel: RelNode): RelNode = { |
| val logicalMatch = rel.asInstanceOf[LogicalMatch] |
| val traitSet = rel.getTraitSet.replace(FlinkConventions.LOGICAL) |
| val newInput = RelOptRule.convert(logicalMatch.getInput, FlinkConventions.LOGICAL) |
| |
| new FlinkLogicalMatch( |
| rel.getCluster, |
| traitSet, |
| newInput, |
| logicalMatch.getRowType, |
| logicalMatch.getPattern, |
| logicalMatch.isStrictStart, |
| logicalMatch.isStrictEnd, |
| logicalMatch.getPatternDefinitions, |
| logicalMatch.getMeasures, |
| logicalMatch.getAfter, |
| logicalMatch.getSubsets, |
| logicalMatch.isAllRows, |
| logicalMatch.getPartitionKeys, |
| logicalMatch.getOrderKeys, |
| logicalMatch.getInterval) |
| } |
| } |
| |
| object FlinkLogicalMatch { |
| val CONVERTER: ConverterRule = new FlinkLogicalMatchConverter() |
| } |