-
Notifications
You must be signed in to change notification settings - Fork 1.6k
feat: [datafusion-spark] Implement next_day
function
#16780
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 2 commits
26b6f24
1314e81
0c2b126
2ab891f
9d5f3e8
4d92a7e
e205d7c
3640d95
5c172d4
172b1e1
e83d752
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -15,11 +15,24 @@ | |
// specific language governing permissions and limitations | ||
// under the License. | ||
|
||
pub mod next_day; | ||
|
||
use datafusion_expr::ScalarUDF; | ||
use datafusion_functions::make_udf_function; | ||
use std::sync::Arc; | ||
|
||
pub mod expr_fn {} | ||
make_udf_function!(next_day::SparkNextDay, next_day); | ||
|
||
pub mod expr_fn { | ||
use datafusion_functions::export_functions; | ||
|
||
export_functions!(( | ||
next_day, | ||
"Returns the first date which is later than start_date and named as indicated. The function returns NULL if at least one of the input parameters is NULL. When both of the input parameters are not NULL and day_of_week is an invalid input, the function throws SparkIllegalArgumentException if spark.sql.ansi.enabled is set to true, otherwise NULL.", | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I think this needs to be adjusted. Rust does not have exceptions and ansi mode is not hooked up yet (might need something like #16661 for that to happen) There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I've removed that part of the doc and moved it to a comment for now. |
||
arg1 arg2 | ||
)); | ||
} | ||
|
||
pub fn functions() -> Vec<Arc<ScalarUDF>> { | ||
vec![] | ||
vec![next_day()] | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,255 @@ | ||
// Licensed to the Apache Software Foundation (ASF) under one | ||
// or more contributor license agreements. See the NOTICE file | ||
// distributed with this work for additional information | ||
// regarding copyright ownership. The ASF licenses this file | ||
// to you under the Apache License, Version 2.0 (the | ||
// "License"); you may not use this file except in compliance | ||
// with the License. You may obtain a copy of the License at | ||
// | ||
// http://www.apache.org/licenses/LICENSE-2.0 | ||
// | ||
// Unless required by applicable law or agreed to in writing, | ||
// software distributed under the License is distributed on an | ||
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY | ||
// KIND, either express or implied. See the License for the | ||
// specific language governing permissions and limitations | ||
// under the License. | ||
|
||
use std::any::Any; | ||
use std::sync::Arc; | ||
|
||
use arrow::array::{new_null_array, ArrayRef, AsArray, Date32Array, StringArrayType}; | ||
use arrow::datatypes::{DataType, Date32Type}; | ||
use chrono::{Datelike, Duration, Weekday}; | ||
use datafusion_common::types::NativeType; | ||
use datafusion_common::{exec_err, plan_err, Result, ScalarValue}; | ||
use datafusion_expr::{ | ||
ColumnarValue, ScalarFunctionArgs, ScalarUDFImpl, Signature, Volatility, | ||
}; | ||
|
||
/// <https://spark.apache.org/docs/latest/api/sql/index.html#next_day> | ||
#[derive(Debug)] | ||
pub struct SparkNextDay { | ||
signature: Signature, | ||
} | ||
|
||
impl Default for SparkNextDay { | ||
fn default() -> Self { | ||
Self::new() | ||
} | ||
} | ||
|
||
impl SparkNextDay { | ||
pub fn new() -> Self { | ||
Self { | ||
signature: Signature::user_defined(Volatility::Immutable), | ||
petern48 marked this conversation as resolved.
Show resolved
Hide resolved
|
||
} | ||
} | ||
} | ||
|
||
impl ScalarUDFImpl for SparkNextDay { | ||
fn as_any(&self) -> &dyn Any { | ||
self | ||
} | ||
|
||
fn name(&self) -> &str { | ||
"next_day" | ||
} | ||
|
||
fn signature(&self) -> &Signature { | ||
&self.signature | ||
} | ||
|
||
fn return_type(&self, _arg_types: &[DataType]) -> Result<DataType> { | ||
Ok(DataType::Date32) | ||
} | ||
|
||
fn invoke_with_args(&self, args: ScalarFunctionArgs) -> Result<ColumnarValue> { | ||
let ScalarFunctionArgs { args, .. } = args; | ||
let [date, day_of_week] = args.as_slice() else { | ||
return exec_err!( | ||
"Spark `next_day` function requires 2 arguments, got {}", | ||
args.len() | ||
); | ||
}; | ||
|
||
match (date, day_of_week) { | ||
(ColumnarValue::Scalar(date), ColumnarValue::Scalar(day_of_week)) => { | ||
match (date, day_of_week) { | ||
(ScalarValue::Date32(days), ScalarValue::Utf8(day_of_week) | ScalarValue::LargeUtf8(day_of_week) | ScalarValue::Utf8View(day_of_week)) => { | ||
if let Some(days) = days { | ||
if let Some(day_of_week) = day_of_week { | ||
Ok(ColumnarValue::Scalar(ScalarValue::Date32( | ||
spark_next_day(*days, day_of_week.as_str()), | ||
))) | ||
} else { | ||
// TODO: if spark.sql.ansi.enabled is false, | ||
// returns NULL instead of an error for a malformed dayOfWeek. | ||
Ok(ColumnarValue::Scalar(ScalarValue::Date32(None))) | ||
} | ||
} else { | ||
Ok(ColumnarValue::Scalar(ScalarValue::Date32(None))) | ||
} | ||
} | ||
_ => exec_err!("Spark `next_day` function: first arg must be date, second arg must be string. Got {args:?}"), | ||
} | ||
} | ||
(ColumnarValue::Array(date_array), ColumnarValue::Scalar(day_of_week)) => { | ||
match (date_array.data_type(), day_of_week) { | ||
(DataType::Date32, ScalarValue::Utf8(day_of_week) | ScalarValue::LargeUtf8(day_of_week) | ScalarValue::Utf8View(day_of_week)) => { | ||
if let Some(day_of_week) = day_of_week { | ||
let result: Date32Array = date_array | ||
.as_primitive::<Date32Type>() | ||
.unary_opt(|days| spark_next_day(days, day_of_week.as_str())) | ||
.with_data_type(DataType::Date32); | ||
Ok(ColumnarValue::Array(Arc::new(result) as ArrayRef)) | ||
} else { | ||
// TODO: if spark.sql.ansi.enabled is false, | ||
// returns NULL instead of an error for a malformed dayOfWeek. | ||
Ok(ColumnarValue::Array(Arc::new(new_null_array(&DataType::Date32, date_array.len())))) | ||
} | ||
} | ||
_ => exec_err!("Spark `next_day` function: first arg must be date, second arg must be string. Got {args:?}"), | ||
} | ||
} | ||
( | ||
ColumnarValue::Array(date_array), | ||
ColumnarValue::Array(day_of_week_array), | ||
) => { | ||
let result = match (date_array.data_type(), day_of_week_array.data_type()) | ||
{ | ||
( | ||
DataType::Date32, | ||
DataType::Utf8 | DataType::LargeUtf8 | DataType::Utf8View, | ||
) => { | ||
let date_array: &Date32Array = | ||
date_array.as_primitive::<Date32Type>(); | ||
match day_of_week_array.data_type() { | ||
DataType::Utf8 => { | ||
let day_of_week_array = | ||
day_of_week_array.as_string::<i32>(); | ||
process_next_day_arrays(date_array, day_of_week_array) | ||
} | ||
DataType::LargeUtf8 => { | ||
let day_of_week_array = | ||
day_of_week_array.as_string::<i64>(); | ||
process_next_day_arrays(date_array, day_of_week_array) | ||
} | ||
DataType::Utf8View => { | ||
let day_of_week_array = | ||
day_of_week_array.as_string_view(); | ||
process_next_day_arrays(date_array, day_of_week_array) | ||
} | ||
other => { | ||
exec_err!("Spark `next_day` function: second arg must be string. Got {other:?}") | ||
} | ||
} | ||
} | ||
(left, right) => { | ||
exec_err!( | ||
"Spark `next_day` function: first arg must be date, second arg must be string. Got {left:?}, {right:?}" | ||
) | ||
} | ||
}?; | ||
Ok(ColumnarValue::Array(result)) | ||
} | ||
_ => exec_err!("Unsupported args {args:?} for Spark function `next_day`"), | ||
} | ||
} | ||
|
||
fn coerce_types(&self, arg_types: &[DataType]) -> Result<Vec<DataType>> { | ||
if arg_types.len() != 2 { | ||
return exec_err!( | ||
"Spark `next_day` function requires 2 arguments, got {}", | ||
arg_types.len() | ||
); | ||
} | ||
|
||
let current_native_type: NativeType = (&arg_types[0]).into(); | ||
if matches!(current_native_type, NativeType::Date) | ||
|| matches!(current_native_type, NativeType::String) | ||
|| matches!(current_native_type, NativeType::Null) | ||
{ | ||
if matches!(&arg_types[1], DataType::Utf8) | ||
|| matches!(&arg_types[1], DataType::LargeUtf8) | ||
|| matches!(&arg_types[1], DataType::Utf8View) | ||
{ | ||
Ok(vec![DataType::Date32, arg_types[1].clone()]) | ||
} else { | ||
plan_err!( | ||
"The second argument of the Spark `next_day` function must be a string, but got {}", | ||
&arg_types[1] | ||
) | ||
} | ||
} else { | ||
plan_err!( | ||
"The first argument of the Spark `next_day` function can only be a date or string, but got {}", &arg_types[0] | ||
) | ||
} | ||
} | ||
} | ||
|
||
fn process_next_day_arrays<'a, S>( | ||
date_array: &Date32Array, | ||
day_of_week_array: &'a S, | ||
) -> Result<ArrayRef> | ||
where | ||
&'a S: StringArrayType<'a>, | ||
{ | ||
let result = date_array | ||
.iter() | ||
.zip(day_of_week_array.iter()) | ||
.map(|(days, day_of_week)| { | ||
if let Some(days) = days { | ||
if let Some(day_of_week) = day_of_week { | ||
spark_next_day(days, day_of_week) | ||
} else { | ||
// TODO: if spark.sql.ansi.enabled is false, | ||
// returns NULL instead of an error for a malformed dayOfWeek. | ||
None | ||
} | ||
} else { | ||
None | ||
} | ||
}) | ||
.collect::<Date32Array>(); | ||
Ok(Arc::new(result) as ArrayRef) | ||
} | ||
|
||
fn spark_next_day(days: i32, day_of_week: &str) -> Option<i32> { | ||
let date = Date32Type::to_naive_date(days); | ||
|
||
petern48 marked this conversation as resolved.
Show resolved
Hide resolved
|
||
let day_of_week = day_of_week.trim().to_uppercase(); | ||
let day_of_week = match day_of_week.as_str() { | ||
"MO" | "MON" | "MONDAY" => Some("MONDAY"), | ||
"TU" | "TUE" | "TUESDAY" => Some("TUESDAY"), | ||
"WE" | "WED" | "WEDNESDAY" => Some("WEDNESDAY"), | ||
"TH" | "THU" | "THURSDAY" => Some("THURSDAY"), | ||
"FR" | "FRI" | "FRIDAY" => Some("FRIDAY"), | ||
"SA" | "SAT" | "SATURDAY" => Some("SATURDAY"), | ||
"SU" | "SUN" | "SUNDAY" => Some("SUNDAY"), | ||
_ => { | ||
// TODO: if spark.sql.ansi.enabled is false, | ||
// returns NULL instead of an error for a malformed dayOfWeek. | ||
None | ||
} | ||
}; | ||
|
||
if let Some(day_of_week) = day_of_week { | ||
let day_of_week = day_of_week.parse::<Weekday>(); | ||
match day_of_week { | ||
Ok(day_of_week) => Some(Date32Type::from_naive_date( | ||
date + Duration::days( | ||
(7 - date.weekday().days_since(day_of_week)) as i64, | ||
), | ||
)), | ||
Err(_) => { | ||
// TODO: if spark.sql.ansi.enabled is false, | ||
// returns NULL instead of an error for a malformed dayOfWeek. | ||
None | ||
} | ||
} | ||
} else { | ||
None | ||
} | ||
} |
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -23,5 +23,17 @@ | |
|
||
## Original Query: SELECT next_day('2015-01-14', 'TU'); | ||
## PySpark 3.5.5 Result: {'next_day(2015-01-14, TU)': datetime.date(2015, 1, 20), 'typeof(next_day(2015-01-14, TU))': 'date', 'typeof(2015-01-14)': 'string', 'typeof(TU)': 'string'} | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. You can safely remove the commented-out tests, since their functionality is already covered below. There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Done, thanks |
||
#query | ||
#SELECT next_day('2015-01-14'::string, 'TU'::string); | ||
query D | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I recommend to add tests for invalid inputs:
Also here we only checked There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Added all of these cases |
||
SELECT next_day('2015-01-14'::string, 'TU'::string); | ||
---- | ||
2015-01-20 | ||
|
||
query D | ||
SELECT next_day('2015-07-27'::string, 'Sun'::string); | ||
---- | ||
2015-08-02 | ||
|
||
query D | ||
SELECT next_day('2015-07-27'::string, 'Sat'::string); | ||
---- | ||
2015-08-01 |
Uh oh!
There was an error while loading. Please reload this page.