morpheus icon indicating copy to clipboard operation
morpheus copied to clipboard

Feature "ColumnNameAcceptance": Scenario "Keeping used expression 4" fails on Spark 3.0

Open DarthMax opened this issue 5 years ago • 0 comments

cannot resolve 'CAST(avg(CAST(NULL AS DOUBLE)) AS NULL)' due to data type mismatch: cannot cast double to null;;
'Aggregate [cast(avg(cast(null as double)) as null) AS   FRESHID27 __ NULL#48721]
+- Join Inner, (target(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48685 = b __ NODE @ session_tmp190#48703)
   :- Join Inner, (node __ NODE @ session_tmp190#48689 = source(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48684)
   :  :- Project [___id#48681 AS node __ NODE @ session_tmp190#48689]
   :  :  +- Project [cast(cast(___id#48678L as string) as binary) AS ___id#48681]
   :  :     +- Project [___id#48676L AS ___id#48678L]
   :  :        +- LocalRelation [___id#48676L]
   :  +- LocalRelation <empty>, [rel __ RELATIONSHIP @ session_tmp190#48683, source(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48684, target(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48685]
   +- Project [node __ NODE @ session_tmp190#48689 AS b __ NODE @ session_tmp190#48703]
      +- Project [___id#48681 AS node __ NODE @ session_tmp190#48689]
         +- Project [cast(cast(___id#48678L as string) as binary) AS ___id#48681]
            +- Project [___id#48676L AS ___id#48678L]
               +- LocalRelation [___id#48676L]

org.apache.spark.sql.AnalysisException: cannot resolve 'CAST(avg(CAST(NULL AS DOUBLE)) AS NULL)' due to data type mismatch: cannot cast double to null;;
'Aggregate [cast(avg(cast(null as double)) as null) AS   FRESHID27 __ NULL#48721]
+- Join Inner, (target(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48685 = b __ NODE @ session_tmp190#48703)
   :- Join Inner, (node __ NODE @ session_tmp190#48689 = source(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48684)
   :  :- Project [___id#48681 AS node __ NODE @ session_tmp190#48689]
   :  :  +- Project [cast(cast(___id#48678L as string) as binary) AS ___id#48681]
   :  :     +- Project [___id#48676L AS ___id#48678L]
   :  :        +- LocalRelation [___id#48676L]
   :  +- LocalRelation <empty>, [rel __ RELATIONSHIP @ session_tmp190#48683, source(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48684, target(rel __ RELATIONSHIP @ session_tmp190) __ NODE#48685]
   +- Project [node __ NODE @ session_tmp190#48689 AS b __ NODE @ session_tmp190#48703]
      +- Project [___id#48681 AS node __ NODE @ session_tmp190#48689]
         +- Project [cast(cast(___id#48678L as string) as binary) AS ___id#48681]
            +- Project [___id#48676L AS ___id#48678L]
               +- LocalRelation [___id#48676L]

	at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42)
	at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$$nestedInanonfun$checkAnalysis$1$2.applyOrElse(CheckAnalysis.scala:117)
	at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$$nestedInanonfun$checkAnalysis$1$2.applyOrElse(CheckAnalysis.scala:108)
	at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformUp$2(TreeNode.scala:280)
	at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:72)
	at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:280)
	at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformUp$1(TreeNode.scala:277)
	at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$mapChildren$1(TreeNode.scala:328)
	at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:189)
	at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:326)
	at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:277)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.$anonfun$transformExpressionsUp$1(QueryPlan.scala:94)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.$anonfun$mapExpressions$1(QueryPlan.scala:106)
	at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:72)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpression$1(QueryPlan.scala:106)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.recursiveTransform$1(QueryPlan.scala:117)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.$anonfun$mapExpressions$3(QueryPlan.scala:122)
	at scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:237)
	at scala.collection.immutable.List.foreach(List.scala:392)
	at scala.collection.TraversableLike.map(TraversableLike.scala:237)
	at scala.collection.TraversableLike.map$(TraversableLike.scala:230)
	at scala.collection.immutable.List.map(List.scala:298)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.recursiveTransform$1(QueryPlan.scala:122)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.$anonfun$mapExpressions$4(QueryPlan.scala:127)
	at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:189)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.mapExpressions(QueryPlan.scala:127)
	at org.apache.spark.sql.catalyst.plans.QueryPlan.transformExpressionsUp(QueryPlan.scala:94)
	at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$1(CheckAnalysis.scala:108)
	at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$1$adapted(CheckAnalysis.scala:86)
	at org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:129)
	at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis(CheckAnalysis.scala:86)
	at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis$(CheckAnalysis.scala:83)
	at org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis(Analyzer.scala:95)
	at org.apache.spark.sql.catalyst.analysis.Analyzer.$anonfun$executeAndCheck$1(Analyzer.scala:109)
	at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:201)
	at org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:106)
	at org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:62)
	at org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111)
	at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:60)
	at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:60)
	at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:52)
	at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:79)
	at org.apache.spark.sql.RelationalGroupedDataset.toDF(RelationalGroupedDataset.scala:65)
	at org.apache.spark.sql.RelationalGroupedDataset.agg(RelationalGroupedDataset.scala:224)
	at org.apache.spark.sql.Dataset.agg(Dataset.scala:1810)
	at org.apache.spark.graph.cypher.SparkTable$DataFrameTable.$anonfun$group$12(SparkTable.scala:167)
	at scala.util.Either.fold(Either.scala:191)
	at org.apache.spark.graph.cypher.SparkTable$DataFrameTable.group(SparkTable.scala:167)
	at org.apache.spark.graph.cypher.SparkTable$DataFrameTable.group(SparkTable.scala:21)
	at org.opencypher.okapi.relational.impl.operators.Aggregate._table$lzycompute(RelationalOperator.scala:346)
	at org.opencypher.okapi.relational.impl.operators.Aggregate._table(RelationalOperator.scala:344)
	at org.opencypher.okapi.relational.impl.operators.RelationalOperator.table(RelationalOperator.scala:77)
	at org.opencypher.okapi.relational.impl.operators.RelationalOperator._table(RelationalOperator.scala:60)
	at org.opencypher.okapi.relational.impl.operators.RelationalOperator.table(RelationalOperator.scala:77)
	at org.opencypher.okapi.relational.impl.operators.Select._table$lzycompute(RelationalOperator.scala:320)
	at org.opencypher.okapi.relational.impl.operators.Select._table(RelationalOperator.scala:317)
	at org.opencypher.okapi.relational.impl.operators.RelationalOperator.table(RelationalOperator.scala:77)
	at org.opencypher.okapi.relational.impl.operators.Select._table$lzycompute(RelationalOperator.scala:320)
	at org.opencypher.okapi.relational.impl.operators.Select._table(RelationalOperator.scala:317)
	at org.opencypher.okapi.relational.impl.operators.RelationalOperator.table(RelationalOperator.scala:77)
	at org.opencypher.okapi.relational.api.planning.RelationalCypherResult.$anonfun$getRecords$1(RelationalCypherResult.scala:71)
	at scala.Option.flatMap(Option.scala:188)
	at org.opencypher.okapi.relational.api.planning.RelationalCypherResult.getRecords(RelationalCypherResult.scala:65)
	at org.opencypher.okapi.api.graph.CypherResult.records(CypherResult.scala:71)
	at org.opencypher.okapi.api.graph.CypherResult.records$(CypherResult.scala:71)
	at org.opencypher.okapi.relational.api.planning.RelationalCypherResult.records(RelationalCypherResult.scala:39)
	at org.opencypher.okapi.tck.test.TCKGraph.execute(TCKFixture.scala:96)
	at org.opencypher.tools.tck.api.Scenario$ScenarioExecutionContext.execute(Scenario.scala:168)
	at org.opencypher.tools.tck.api.Scenario.$anonfun$executeOnGraph$1(Scenario.scala:66)
	at scala.collection.LinearSeqOptimized.foldLeft(LinearSeqOptimized.scala:126)
	at scala.collection.LinearSeqOptimized.foldLeft$(LinearSeqOptimized.scala:122)
	at scala.collection.immutable.List.foldLeft(List.scala:89)
	at org.opencypher.tools.tck.api.Scenario.executeOnGraph(Scenario.scala:60)
	at org.opencypher.tools.tck.api.Scenario$$anon$1.execute(Scenario.scala:54)
	at org.apache.spark.graph.cypher.tck.SparkCypherTckSuite.$anonfun$new$2(SparkCypherTckSuite.scala:32)
	at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
	at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
	at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
	at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
	at org.scalatest.Transformer.apply(Transformer.scala:22)
	at org.scalatest.Transformer.apply(Transformer.scala:20)
	at org.scalatest.FunSuiteLike$$anon$1.apply(FunSuiteLike.scala:186)
	at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:104)
	at org.scalatest.FunSuiteLike.invokeWithFixture$1(FunSuiteLike.scala:184)
	at org.scalatest.FunSuiteLike.$anonfun$runTest$1(FunSuiteLike.scala:196)
	at org.scalatest.SuperEngine.runTestImpl(Engine.scala:289)
	at org.scalatest.FunSuiteLike.runTest(FunSuiteLike.scala:196)
	at org.scalatest.FunSuiteLike.runTest$(FunSuiteLike.scala:178)
	at org.apache.spark.graph.cypher.tck.SparkCypherTckSuite.org$scalatest$BeforeAndAfterEach$$super$runTest(SparkCypherTckSuite.scala:18)
	at org.scalatest.BeforeAndAfterEach.runTest(BeforeAndAfterEach.scala:221)
	at org.scalatest.BeforeAndAfterEach.runTest$(BeforeAndAfterEach.scala:214)
	at org.apache.spark.graph.cypher.tck.SparkCypherTckSuite.runTest(SparkCypherTckSuite.scala:18)
	at org.scalatest.FunSuiteLike.$anonfun$runTests$1(FunSuiteLike.scala:229)
	at org.scalatest.SuperEngine.$anonfun$runTestsInBranch$1(Engine.scala:396)
	at scala.collection.immutable.List.foreach(List.scala:392)
	at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:384)
	at org.scalatest.SuperEngine.runTestsInBranch(Engine.scala:379)
	at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:461)
	at org.scalatest.FunSuiteLike.runTests(FunSuiteLike.scala:229)
	at org.scalatest.FunSuiteLike.runTests$(FunSuiteLike.scala:228)
	at org.scalatest.FunSuite.runTests(FunSuite.scala:1560)
	at org.scalatest.Suite.run(Suite.scala:1147)
	at org.scalatest.Suite.run$(Suite.scala:1129)
	at org.scalatest.FunSuite.org$scalatest$FunSuiteLike$$super$run(FunSuite.scala:1560)
	at org.scalatest.FunSuiteLike.$anonfun$run$1(FunSuiteLike.scala:233)
	at org.scalatest.SuperEngine.runImpl(Engine.scala:521)
	at org.scalatest.FunSuiteLike.run(FunSuiteLike.scala:233)
	at org.scalatest.FunSuiteLike.run$(FunSuiteLike.scala:232)
	at org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:53)
	at org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:213)
	at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
	at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
	at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:53)
	at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:45)
	at org.scalatest.tools.Runner$.$anonfun$doRunRunRunDaDoRunRun$13(Runner.scala:1340)
	at org.scalatest.tools.Runner$.$anonfun$doRunRunRunDaDoRunRun$13$adapted(Runner.scala:1334)
	at scala.collection.immutable.List.foreach(List.scala:392)
	at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:1334)
	at org.scalatest.tools.Runner$.$anonfun$runOptionallyWithPassFailReporter$24(Runner.scala:1031)
	at org.scalatest.tools.Runner$.$anonfun$runOptionallyWithPassFailReporter$24$adapted(Runner.scala:1010)
	at org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:1500)
	at org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:1010)
	at org.scalatest.tools.Runner$.run(Runner.scala:850)
	at org.scalatest.tools.Runner.run(Runner.scala)
	at org.jetbrains.plugins.scala.testingSupport.scalaTest.ScalaTestRunner.runScalaTest2(ScalaTestRunner.java:131)
	at org.jetbrains.plugins.scala.testingSupport.scalaTest.ScalaTestRunner.main(ScalaTestRunner.java:28)

DarthMax avatar Mar 12 '19 17:03 DarthMax