Databricks-Spark-Rules
Scalafix rules for Databricks Spark.
Databricks provides managed Spark clusters and they use SparkContext and SparkSession from runtime.
Developers must be careful not to call unsafe methods like SparkContext.stop or System.exit. This Scalafix rules lint ocurrences of such methods.
I planed to add rules for performance tuning in the future.
| pre-release | release |
|---|---|
Install
sbt
ThisBuild / scalafixDependencies += "dev.i10416" %% "databricks-spark-rules" % "0.0.1"Mill
def scalafixIvyDeps = Agg(ivy"dev.i10416::databricks-spark-rules:0.0.1")How to Use
| rule | description |
|---|---|
| UnsafeSparkStopCall | warn SparkContext#stop and SparkSession#stop |
| UnsafeSparkConstructors | warn new SparkContext |
| UnsafeSysExit | warn System.exit(0) |
How to contribute?
- Give it a star
⭐ - Drop the feedback to the author @i10416
- Send a PR with fixes of typos/bugs/etc
🐛
License
Licensed under the Apache License, Version 2.0.