blob: 054da4bedfa79e29dee065634a4e5f01e68778d1 [file] [log] [blame]
package org.apache.spark.sql.hive
import org.apache.spark.sql.CarbonDatasourceHadoopRelation
import org.apache.spark.sql.catalyst.expressions.{Exists, In, ListQuery, ScalarSubquery}
import org.apache.spark.sql.catalyst.plans.logical.{Filter, LogicalPlan}
import org.apache.spark.sql.execution.datasources.LogicalRelation
object CarbonOptimizerUtil {
def transformForScalarSubQuery(plan: LogicalPlan): LogicalPlan = {
// In case scalar subquery add flag in relation to skip the decoder plan in optimizer rule, And
// optimize whole plan at once.
val transFormedPlan = plan.transform {
case filter: Filter =>
filter.transformExpressions {
case s: ScalarSubquery =>
val tPlan = s.plan.transform {
case lr: LogicalRelation
if lr.relation.isInstanceOf[CarbonDatasourceHadoopRelation] =>
lr.relation.asInstanceOf[CarbonDatasourceHadoopRelation].isSubquery += true
lr
}
ScalarSubquery(tPlan, s.children, s.exprId)
case e: Exists =>
val tPlan = e.plan.transform {
case lr: LogicalRelation
if lr.relation.isInstanceOf[CarbonDatasourceHadoopRelation] =>
lr.relation.asInstanceOf[CarbonDatasourceHadoopRelation].isSubquery += true
lr
}
Exists(tPlan, e.children.map(_.canonicalized), e.exprId)
case In(value, Seq(l:ListQuery)) =>
val tPlan = l.plan.transform {
case lr: LogicalRelation
if lr.relation.isInstanceOf[CarbonDatasourceHadoopRelation] =>
lr.relation.asInstanceOf[CarbonDatasourceHadoopRelation].isSubquery += true
lr
}
In(value, Seq(ListQuery(tPlan, l.children, l.exprId)))
}
}
transFormedPlan
}
}