- Spill — when you're out of execution memory
- Excessive retries — when the data is recomputed all over
- Unused executors — when the driver is doing too much
- Several more now, even more coming in future
Automatically detect performance issues in all your Spark jobs
... and not just data skew
Got thousands of jobs? You can use Spark Advisor to automatically find performance regressions as data changes
One job misbehaves? Run it with additional Spark options and we can probably suggest what's wrong
And no marketing calls either