You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Running TPC-H SF=100 on a single node repeatedly will eventually run out of disk when native or auto shuffle mode is enabled. The shuffle files generated when running the queries never gets deleted. Setting spark.cleaner.periodicGC.interval=60s or manually trigger driver GC does not help.
This problem only happens when spark.comet.exec.shuffle.mode is native or auto. It does not happen when shuffle mode is jvm.
Steps to reproduce
Run tpcbench.py with --iterations 10 will take hundreds of gigs of disk space. Here is an example:
Describe the bug
Running TPC-H SF=100 on a single node repeatedly will eventually run out of disk when native or auto shuffle mode is enabled. The shuffle files generated when running the queries never gets deleted. Setting
spark.cleaner.periodicGC.interval=60s
or manually trigger driver GC does not help.This problem only happens when
spark.comet.exec.shuffle.mode
isnative
orauto
. It does not happen when shuffle mode isjvm
.Steps to reproduce
Run tpcbench.py with
--iterations 10
will take hundreds of gigs of disk space. Here is an example:Expected behavior
Unused shuffle files should be deleted when GC was triggered on Spark driver.
Additional context
No response
The text was updated successfully, but these errors were encountered: