Commit a802cfe
committed
Fix materialization when running on Spark cluster.
When running materialization and have Spark offline store configured to use cluster (`spark.master` pointing to actual Spark master node) `self.to_spark_df().write.parquet(temp_dir, mode="overwrite")` will create parquet file in worker node but `return pq.read_table(temp_dir)` is executed on driver node and it can't read from worker. Proposed fix makes materialization work when run on Spark cluster.
Signed-off-by: ckarwicki <104110169+ckarwicki-deloitte@users.noreply.github.com>
Signed-off-by: ckarwicki <71740096+ckarwicki@users.noreply.github.com>1 parent d7b0c52 commit a802cfe
File tree
1 file changed
+2
-5
lines changed- sdk/python/feast/infra/offline_stores/contrib/spark_offline_store
1 file changed
+2
-5
lines changedLines changed: 2 additions & 5 deletions
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
325 | 325 | | |
326 | 326 | | |
327 | 327 | | |
328 | | - | |
329 | | - | |
330 | | - | |
331 | | - | |
332 | | - | |
| 328 | + | |
| 329 | + | |
333 | 330 | | |
334 | 331 | | |
335 | 332 | | |
| |||
0 commit comments