ch.cern.DemoPlugin.scala Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of spark-plugins_2.13 Show documentation
Show all versions of spark-plugins_2.13 Show documentation
SparkPlugins provides code and examples of how to deploy Apache Spark Plugins. Notably focosuing on extensions to the metrics system applied to measuring I/O from cloud Filesystems, OS/system metrics and custom metrics.
The newest version!
package ch.cern
import scala.jdk.CollectionConverters._
import java.util.{Map => JMap}
import org.apache.spark.api.plugin.{DriverPlugin, ExecutorPlugin, PluginContext, SparkPlugin}
import org.apache.spark.SparkContext
// Basic example of Spark Executor Plugin in Scala
class DemoPlugin extends SparkPlugin {
// Return the plugin's driver-side component.
override def driverPlugin(): DriverPlugin = {
new DriverPlugin() {
override def init(sc: SparkContext, myContext: PluginContext): JMap[String, String] = {
DemoPlugin.numSuccessfulPlugins += 1
Map.empty[String, String].asJava
}
override def shutdown(): Unit = {
DemoPlugin.numSuccessfulTerminations += 1
}
}
}
// Return the plugin's executor-side component.
override def executorPlugin(): ExecutorPlugin = {
new ExecutorPlugin() {
override def init(myContext: PluginContext, extraConf: JMap[String, String]): Unit = {
DemoPlugin.numSuccessfulPlugins += 1
}
override def shutdown(): Unit = {
DemoPlugin.numSuccessfulTerminations += 1
}
}
}
}
object DemoPlugin {
var numSuccessfulPlugins : Int = 0
var numSuccessfulTerminations: Int = 0
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy