Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is Databricks auto-scaling supposed to work with joblib-spark? #45

Open
pauljohn32 opened this issue Jan 5, 2023 · 1 comment
Open

Comments

@pauljohn32
Copy link

joblib-spark will use the nodes that are on/available when cluster is started, but it never triggers the Databricks system to wake up another node, even when the one it is using is running on all cores.

I found some comments from 2020 that say auto-scaling with Spark is generally problematic, compared to auto-scaling in Azure itself. So maybe auto-scaling is not supposed to work.

Should it work?

@kashishsehgal73
Copy link

looking at the source code, https://github.com/joblib/joblib-spark/blob/master/joblibspark/backend.py

if we pass n_jobs = -1,

in line 112, it reset the njobs from the spark context object return value in line 128 (which returns the number of active cores in the cluster)
it does not ask for more workers at all.....

So databricks would not get any signal or demand to autoscale,In short, joblibspark is not designed to autoscale databricks clusters at all.

But we should push for this feature as most spark loads are being moved to the cloud and autoscaling is integral part of cost saving and effiiciency

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants