Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -251,7 +251,9 @@ case class ResolvedNonPersistentFunc(
case class ResolvedIdentifier(
catalog: CatalogPlugin,
identifier: Identifier,
override val output: Seq[Attribute] = Nil) extends LeafNodeWithoutStats
override val output: Seq[Attribute] = Nil) extends LeafNodeWithoutStats {
def name: String = (catalog.name +: identifier.namespace :+ identifier.name).quoted
}

object ResolvedIdentifier {
def unapply(ri: ResolvedIdentifier): Option[(CatalogPlugin, Identifier)] = {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -248,7 +248,9 @@ case class DefaultValueExpression(
final override val nodePatterns: Seq[TreePattern] = Seq(ANALYSIS_AWARE_EXPRESSION)

override def dataType: DataType = child.dataType
override def stringArgs: Iterator[Any] = Iterator(child, originalSQL)
override def prettyName: String = "default"
override def sql: String = s"DEFAULT $originalSQL"
override def stringArgs: Iterator[Any] = Iterator(child, s"sql='$originalSQL'")
override def markAsAnalyzed(): DefaultValueExpression =
copy(analyzedChild = Some(child))
override protected def withNewChildInternal(newChild: Expression): Expression =
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,9 @@ case class CreateVariableExec(
defaultExpr: DefaultValueExpression,
replace: Boolean) extends LeafV2CommandExec with ExpressionsEvaluator {

override def stringArgs: Iterator[Any] =
Iterator(resolvedIdentifiers.map(_.name), defaultExpr, s"replace=$replace")

override protected def run(): Seq[InternalRow] = {
val scriptingVariableManager = SqlScriptingContextManager.get().flatMap(_.getVariableManager)
val tempVariableManager = session.sessionState.catalogManager.tempVariableManager
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -33,7 +33,7 @@ CreateDataSourceTableCommand `spark_catalog`.`default`.`x`, false
-- !query
DECLARE sql_string STRING
-- !query analysis
CreateVariable defaultvalueexpression(null, null), false
CreateVariable default(null, sql='null'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.sql_string


Expand Down Expand Up @@ -126,7 +126,7 @@ SetVariable [variablereference(system.session.sql_string='SELECT * from tbl_view
-- !query
DECLARE a STRING
-- !query analysis
CreateVariable defaultvalueexpression(null, null), false
CreateVariable default(null, sql='null'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.a


Expand Down Expand Up @@ -231,7 +231,7 @@ SetVariable [variablereference(system.session.sql_string='SELECT * from tbl_view
-- !query
DECLARE b INT
-- !query analysis
CreateVariable defaultvalueexpression(null, null), false
CreateVariable default(null, sql='null'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.b


Expand Down Expand Up @@ -327,7 +327,7 @@ Project [variablereference(system.session.sql_string='SELECT id from tbl_view wh
-- !query
DECLARE res_id INT
-- !query analysis
CreateVariable defaultvalueexpression(null, null), false
CreateVariable default(null, sql='null'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.res_id


Expand Down Expand Up @@ -507,7 +507,7 @@ org.apache.spark.sql.AnalysisException
-- !query
DECLARE OR REPLACE testvarA INT
-- !query analysis
CreateVariable defaultvalueexpression(null, null), true
CreateVariable default(null, sql='null'), true
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.testvarA


Expand Down Expand Up @@ -782,7 +782,7 @@ org.apache.spark.sql.catalyst.parser.ParseException
-- !query
DECLARE p = 10
-- !query analysis
CreateVariable defaultvalueexpression(10, 10), false
CreateVariable default(10, sql='10'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.p


Expand Down Expand Up @@ -924,7 +924,7 @@ org.apache.spark.sql.AnalysisException
-- !query
DECLARE int_var INT
-- !query analysis
CreateVariable defaultvalueexpression(null, null), false
CreateVariable default(null, sql='null'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.int_var


Expand Down Expand Up @@ -952,7 +952,7 @@ org.apache.spark.sql.AnalysisException
-- !query
DECLARE null_var STRING
-- !query analysis
CreateVariable defaultvalueexpression(null, null), false
CreateVariable default(null, sql='null'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.null_var


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -571,7 +571,7 @@ DropFunctionCommand myDoubleAvg, false, true
-- !query
DECLARE var = 'sometable'
-- !query analysis
CreateVariable defaultvalueexpression(sometable, 'sometable'), false
CreateVariable default(sometable, sql=''sometable''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.var


Expand Down Expand Up @@ -1115,21 +1115,21 @@ DropTable false, false
-- !query
DECLARE agg = 'max'
-- !query analysis
CreateVariable defaultvalueexpression(max, 'max'), false
CreateVariable default(max, sql=''max''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.agg


-- !query
DECLARE col = 'c1'
-- !query analysis
CreateVariable defaultvalueexpression(c1, 'c1'), false
CreateVariable default(c1, sql=''c1''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.col


-- !query
DECLARE tab = 'T'
-- !query analysis
CreateVariable defaultvalueexpression(T, 'T'), false
CreateVariable default(T, sql=''T''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.tab


Expand Down Expand Up @@ -1805,7 +1805,7 @@ DropTable false, false
-- !query
DECLARE IDENTIFIER('my_var') = 'value'
-- !query analysis
CreateVariable defaultvalueexpression(value, 'value'), false
CreateVariable default(value, sql=''value''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.my_var


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -571,7 +571,7 @@ DropFunctionCommand myDoubleAvg, false, true
-- !query
DECLARE var = 'sometable'
-- !query analysis
CreateVariable defaultvalueexpression(sometable, 'sometable'), false
CreateVariable default(sometable, sql=''sometable''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.var


Expand Down Expand Up @@ -1115,21 +1115,21 @@ DropTable false, false
-- !query
DECLARE agg = 'max'
-- !query analysis
CreateVariable defaultvalueexpression(max, 'max'), false
CreateVariable default(max, sql=''max''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.agg


-- !query
DECLARE col = 'c1'
-- !query analysis
CreateVariable defaultvalueexpression(c1, 'c1'), false
CreateVariable default(c1, sql=''c1''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.col


-- !query
DECLARE tab = 'T'
-- !query analysis
CreateVariable defaultvalueexpression(T, 'T'), false
CreateVariable default(T, sql=''T''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.tab


Expand Down Expand Up @@ -1713,7 +1713,7 @@ DropTable false, false
-- !query
DECLARE IDENTIFIER('my_var') = 'value'
-- !query analysis
CreateVariable defaultvalueexpression(value, 'value'), false
CreateVariable default(value, sql=''value''), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.my_var


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -987,7 +987,7 @@ DropTableCommand `spark_catalog`.`default`.`v1`, false, true, false
-- !query
DECLARE v1 INT
-- !query analysis
CreateVariable defaultvalueexpression(null, null), false
CreateVariable default(null, sql='null'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.v1


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -1115,7 +1115,7 @@ DropTableCommand `spark_catalog`.`default`.`v1`, false, true, false
-- !query
DECLARE v1 INT
-- !query analysis
CreateVariable defaultvalueexpression(null, null), false
CreateVariable default(null, sql='null'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.v1


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -4777,7 +4777,7 @@ Project [x#x, y#x, extend#x, (x + 1)#x, y#x]
-- !query
declare or replace extend int = 5
-- !query analysis
CreateVariable defaultvalueexpression(cast(5 as int), 5), true
CreateVariable default(cast(5 as int), sql='5'), true
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.extend


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -792,6 +792,7 @@ Project [a#x, z#x, b#x, y#x, c#x, x#x, d#x, w#x, dotNET#xL, Java#xL]
+- SubqueryAlias courseSales
+- LocalRelation [course#x, year#x, earnings#x]


-- !query
SELECT pv.year, pv.net, pv.jv FROM (
SELECT year, course, earnings FROM courseSales
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,14 +8,14 @@ CreateDataSourceTableCommand `spark_catalog`.`default`.`t1`, false
-- !query
DECLARE all = 1
-- !query analysis
CreateVariable defaultvalueexpression(1, 1), false
CreateVariable default(1, sql='1'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.all


-- !query
DECLARE a = 1
-- !query analysis
CreateVariable defaultvalueexpression(1, 1), false
CreateVariable default(1, sql='1'), false
+- ResolvedIdentifier org.apache.spark.sql.catalyst.analysis.FakeSystemCatalog$@xxxxxxxx, session.a


Expand Down
Loading