Github user jiangxb1987 commented on a diff in the pull request:

    https://github.com/apache/spark/pull/16561#discussion_r96331626
  
    --- Diff: 
sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/view.scala 
---
    @@ -28,22 +28,60 @@ import org.apache.spark.sql.catalyst.rules.Rule
      */
     
     /**
    - * Make sure that a view's child plan produces the view's output 
attributes. We wrap the child
    - * with a Project and add an alias for each output attribute. The 
attributes are resolved by
    - * name. This should be only done after the batch of Resolution, because 
the view attributes are
    - * not completely resolved during the batch of Resolution.
    + * Make sure that a view's child plan produces the view's output 
attributes. We try to wrap the
    + * child by:
    + * 1. Generate the `queryOutput` by:
    + *    1.1. If the query column names are defined, map the column names to 
attributes in the child
    + *         output by name(This is mostly for handling view queries like 
SELECT * FROM ..., the
    + *         schema of the referenced table/view may change after the view 
has been created, so we
    + *         have to save the output of the query to `viewQueryColumnNames`, 
and restore them during
    + *         view resolution, in this way, we are able to get the correct 
view column ordering and
    + *         omit the extra columns that we don't require);
    + *    1.2. Else set the child output attributes to `queryOutput`.
    + * 2. Map the `queryQutput` to view output by index, if the corresponding 
attributes don't match,
    + *    try to up cast and alias the attribute in `queryOutput` to the 
attribute in the view output.
    + * 3. Add a Project over the child, with the new output generated by the 
previous steps.
    + * If the view output doesn't have the same number of columns neither with 
the child output, nor
    + * with the query column names, throw an AnalysisException.
    + *
    + * This should be only done after the batch of Resolution, because the 
view attributes are not
    + * completely resolved during the batch of Resolution.
      */
     case class AliasViewChild(conf: CatalystConf) extends Rule[LogicalPlan] {
       override def apply(plan: LogicalPlan): LogicalPlan = plan 
resolveOperators {
    -    case v @ View(_, output, child) if child.resolved =>
    +    case v @ View(desc, output, child) if child.resolved && output != 
child.output =>
           val resolver = conf.resolver
    -      val newOutput = output.map { attr =>
    -        val originAttr = findAttributeByName(attr.name, child.output, 
resolver)
    -        // The dataType of the output attributes may be not the same with 
that of the view output,
    -        // so we should cast the attribute to the dataType of the view 
output attribute. If the
    -        // cast can't perform, will throw an AnalysisException.
    -        Alias(Cast(originAttr, attr.dataType), attr.name)(exprId = 
attr.exprId,
    -          qualifier = attr.qualifier, explicitMetadata = 
Some(attr.metadata))
    +      val queryColumnNames = desc.viewQueryColumnNames
    +      val queryOutput = if (queryColumnNames.nonEmpty) {
    +        // If the view output doesn't have the same number of columns with 
the query column names,
    +        // throw an AnalysisException.
    +        if (output.length != queryColumnNames.length) {
    +          throw new AnalysisException(
    +            s"The view output ${output.mkString("[", ",", "]")} doesn't 
have the same number of " +
    +              s"columns with the query column names 
${queryColumnNames.mkString("[", ",", "]")}")
    +        }
    +        desc.viewQueryColumnNames.map { colName =>
    +          findAttributeByName(colName, child.output, resolver)
    +        }
    +      } else {
    +        // For view created before Spark 2.2.0, the view text is already 
fully qualified, the plan
    +        // output is the same with the view output.
    +        child.output
    +      }
    +      // Map the attributes in the query output to the attributes in the 
view output by index.
    +      val newOutput = output.zip(queryOutput).map {
    --- End diff --
    
    For views created by older versions of Spark, the view text is fully 
qualified, so the output is the same with the view output. Or else we have 
checked that the output have the same length with `queryColumnNames`. So 
perhaps we don't need to check the size of `output` and `queryOutput` here.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to