org.apache.spark.dataflint.api.DataflintSQLMetricsPage.scala Maven / Gradle / Ivy
Go to download
Show more of this group Show more artifacts with this name
Show all versions of spark_2.12 Show documentation
Show all versions of spark_2.12 Show documentation
Open Source Data-Application Performance Monitoring for Apache Spark
The newest version!
package org.apache.spark.dataflint.api
import org.apache.spark.internal.Logging
import org.apache.spark.sql.execution.ui.{SQLAppStatusListener, SQLAppStatusStore, SparkPlanGraph}
import org.apache.spark.ui.{SparkUI, WebUIPage}
import org.json4s.JsonAST.JValue
import org.json4s.{JsonAST, _}
import javax.servlet.http.HttpServletRequest
import scala.xml.Node
class DataflintSQLMetricsPage(ui: SparkUI, sqlListener: () => Option[SQLAppStatusListener])
extends WebUIPage("sqlmetrics") with Logging {
private var sqlListenerCache: Option[SQLAppStatusListener] = None
override def renderJson(request: HttpServletRequest): JsonAST.JValue = {
try {
if (sqlListenerCache.isEmpty) {
sqlListenerCache = sqlListener()
}
val sqlStore = new SQLAppStatusStore(ui.store.store, sqlListenerCache)
val executionId = request.getParameter("executionId")
if (executionId == null) {
return JObject()
}
val executionIdLong = executionId.toLong
val metrics = sqlStore.executionMetrics(executionIdLong)
val isDatabricks = ui.conf.getOption("spark.databricks.clusterUsageTags.cloudProvider").isDefined
val graph = if (isDatabricks) {
val exec = sqlStore.execution(executionIdLong).get
val planVersion = exec.getClass.getMethod("latestVersion").invoke(exec).asInstanceOf[Long]
sqlStore.getClass.getMethods.filter(_.getName == "planGraph").head.invoke(sqlStore, executionIdLong.asInstanceOf[Object], planVersion.asInstanceOf[Object]).asInstanceOf[SparkPlanGraph]
} else
sqlStore.planGraph(executionIdLong)
val nodesMetrics = graph.allNodes.map(node => NodeMetrics(node.id, node.name, node.metrics.map(metric => {
NodeMetric(metric.name, metrics.get(metric.accumulatorId))
}).toSeq))
// filter nodes without metrics
.filter(nodeMetrics => !nodeMetrics.metrics.forall(_.value.isEmpty))
val jValue: JValue = Extraction.decompose(nodesMetrics)(org.json4s.DefaultFormats)
jValue
} catch {
case e: Throwable => {
logError("failed to serve dataflint SQL metrics", e)
JObject()
}
}
}
override def render(request: HttpServletRequest): Seq[Node] = Seq[Node]()
}
© 2015 - 2025 Weber Informatics LLC | Privacy Policy