@@ -52,7 +52,12 @@ def remove_extra_spaces_sql(df, column_name):
5252
5353@pytest .fixture
5454def spark_fixture ():
55- spark = SparkSession .builder .appName ("Testing PySpark Example" ).getOrCreate ()
55+ spark = (
56+ SparkSession .builder .appName ("Testing PySpark Example" )
57+ .config ("spark.driver.host" , "127.0.0.1" )
58+ .config ("spark.driver.bindAddress" , "127.0.0.1" )
59+ .getOrCreate ()
60+ )
5661 try :
5762 yield spark
5863 finally :
@@ -61,7 +66,12 @@ def spark_fixture():
6166
6267@patch ("feast.infra.compute_engines.spark.utils.get_or_create_new_spark_session" )
6368def test_spark_transformation (spark_fixture ):
64- spark = SparkSession .builder .appName ("Testing PySpark Example" ).getOrCreate ()
69+ spark = (
70+ SparkSession .builder .appName ("Testing PySpark Example" )
71+ .config ("spark.driver.host" , "127.0.0.1" )
72+ .config ("spark.driver.bindAddress" , "127.0.0.1" )
73+ .getOrCreate ()
74+ )
6575 df = get_sample_df (spark )
6676
6777 spark_transformation = Transformation (
@@ -77,7 +87,12 @@ def test_spark_transformation(spark_fixture):
7787
7888@patch ("feast.infra.compute_engines.spark.utils.get_or_create_new_spark_session" )
7989def test_spark_transformation_init_transformation (spark_fixture ):
80- spark = SparkSession .builder .appName ("Testing PySpark Example" ).getOrCreate ()
90+ spark = (
91+ SparkSession .builder .appName ("Testing PySpark Example" )
92+ .config ("spark.driver.host" , "127.0.0.1" )
93+ .config ("spark.driver.bindAddress" , "127.0.0.1" )
94+ .getOrCreate ()
95+ )
8196 df = get_sample_df (spark )
8297
8398 spark_transformation = SparkTransformation (
@@ -93,7 +108,12 @@ def test_spark_transformation_init_transformation(spark_fixture):
93108
94109@patch ("feast.infra.compute_engines.spark.utils.get_or_create_new_spark_session" )
95110def test_spark_transformation_sql (spark_fixture ):
96- spark = SparkSession .builder .appName ("Testing PySpark Example" ).getOrCreate ()
111+ spark = (
112+ SparkSession .builder .appName ("Testing PySpark Example" )
113+ .config ("spark.driver.host" , "127.0.0.1" )
114+ .config ("spark.driver.bindAddress" , "127.0.0.1" )
115+ .getOrCreate ()
116+ )
97117 df = get_sample_df (spark )
98118
99119 spark_transformation = SparkTransformation (
0 commit comments