Skip to content

Commit

Permalink
suppress warning message of pandas_on_spark to_spark (#1058)
Browse files Browse the repository at this point in the history
  • Loading branch information
thinkall authored Jun 1, 2023
1 parent b975df8 commit d36b2af
Show file tree
Hide file tree
Showing 2 changed files with 16 additions and 0 deletions.
4 changes: 4 additions & 0 deletions flaml/automl/spark/metrics.py
Original file line number Diff line number Diff line change
Expand Up @@ -54,6 +54,10 @@ def spark_metric_loss_score(
Returns:
float | the loss score. A lower value indicates a better model.
"""
import warnings

warnings.filterwarnings("ignore")

label_col = "label"
prediction_col = "prediction"
kwargs = {}
Expand Down
12 changes: 12 additions & 0 deletions flaml/automl/spark/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -92,6 +92,10 @@ def train_test_split_pyspark(
pyspark.sql.DataFrame/pandas_on_spark DataFrame | The train dataframe.
pyspark.sql.DataFrame/pandas_on_spark DataFrame | The test dataframe.
"""
import warnings

warnings.filterwarnings("ignore")

if isinstance(df, psDataFrame):
df = df.to_spark(index_col=index_col)

Expand Down Expand Up @@ -156,6 +160,10 @@ def iloc_pandas_on_spark(
index_col: Optional[str] = "tmp_index_col",
) -> Union[psDataFrame, psSeries]:
"""Get the rows of a pandas_on_spark dataframe/series by index."""
import warnings

warnings.filterwarnings("ignore")

if isinstance(psdf, (DataFrame, Series)):
return psdf.iloc[index]
if isinstance(index, (int, slice)):
Expand Down Expand Up @@ -207,6 +215,10 @@ def spark_kFold(
Returns:
A list of (train, validation) DataFrames.
"""
import warnings

warnings.filterwarnings("ignore")

if isinstance(dataset, psDataFrame):
dataset = dataset.to_spark(index_col=index_col)

Expand Down

0 comments on commit d36b2af

Please sign in to comment.