Skip to content

Commit 4d88390

Browse files
address comments
1 parent 1621de2 commit 4d88390

File tree

2 files changed

+15
-17
lines changed

2 files changed

+15
-17
lines changed

python/pyspark/sql/tests/arrow/test_arrow_udtf.py

Lines changed: 15 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -482,9 +482,9 @@ def eval(self, input_val: "pa.Array") -> Iterator["pa.Table"]:
482482
with self.assertRaisesRegex(Exception, "LATERAL_JOIN_WITH_ARROW_UDTF_UNSUPPORTED"):
483483
self.spark.sql(
484484
"""
485-
SELECT t.id, f.x, f.result
486-
FROM test_table t, LATERAL simple_arrow_udtf(t.id) f
487-
"""
485+
SELECT t.id, f.x, f.result
486+
FROM test_table t, LATERAL simple_arrow_udtf(t.id) f
487+
"""
488488
)
489489

490490
def test_arrow_udtf_lateral_join_with_table_argument_disallowed(self):
@@ -513,9 +513,9 @@ def eval(self, input_table: "pa.Table") -> Iterator["pa.Table"]:
513513
):
514514
self.spark.sql(
515515
"""
516-
SELECT t1.id, f.filtered_id
517-
FROM test_table1 t1, LATERAL mixed_args_udtf(table(SELECT * FROM test_table2)) f
518-
"""
516+
SELECT t1.id, f.filtered_id
517+
FROM test_table1 t1, LATERAL mixed_args_udtf(table(SELECT * FROM test_table2)) f
518+
"""
519519
)
520520

521521
def test_arrow_udtf_with_table_argument_then_lateral_join_allowed(self):
@@ -536,11 +536,11 @@ def eval(self, input_table: "pa.Table") -> Iterator["pa.Table"]:
536536

537537
result_df = self.spark.sql(
538538
"""
539-
SELECT f.processed_id, j.label
540-
FROM table_arg_udtf(table(SELECT * FROM source_table)) f,
541-
join_table j
542-
ORDER BY f.processed_id, j.label
543-
"""
539+
SELECT f.processed_id, j.label
540+
FROM table_arg_udtf(table(SELECT * FROM source_table)) f,
541+
join_table j
542+
ORDER BY f.processed_id, j.label
543+
"""
544544
)
545545

546546
expected_data = [
@@ -581,10 +581,10 @@ def eval(self, input_val: int):
581581

582582
result_df = self.spark.sql(
583583
"""
584-
SELECT c.computed_value, m.multiplied
585-
FROM compute_udtf(table(SELECT * FROM values_table) WITH SINGLE PARTITION) c,
586-
LATERAL multiply_udtf(c.computed_value) m
587-
"""
584+
SELECT c.computed_value, m.multiplied
585+
FROM compute_udtf(table(SELECT * FROM values_table) WITH SINGLE PARTITION) c,
586+
LATERAL multiply_udtf(c.computed_value) m
587+
"""
588588
)
589589

590590
expected_df = self.spark.createDataFrame([(60, 180)], "computed_value int, multiplied int")

sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Analyzer.scala

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4252,8 +4252,6 @@ object RemoveTempResolvedColumn extends Rule[LogicalPlan] {
42524252
}
42534253
}
42544254

4255-
4256-
42574255
/**
42584256
* Rule that's used to handle `UnresolvedHaving` nodes with resolved `condition` and `child`.
42594257
* It's placed outside the main batch to avoid conflicts with other rules that resolve

0 commit comments

Comments
 (0)