streaming.dsl.mmlib.algs.SQLFPGrowth.scala Maven / Gradle / Ivy
The newest version!
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package streaming.dsl.mmlib.algs
import org.apache.spark.ml.fpm.{FPGrowth, FPGrowthModel}
import org.apache.spark.ml.linalg.SQLDataTypes._
import streaming.dsl.mmlib.SQLAlg
import org.apache.spark.sql.{SparkSession, _}
import org.apache.spark.sql.expressions.UserDefinedFunction
import org.apache.spark.sql.types.{ArrayType, IntegerType, ObjectType, StringType}
/**
* Created by allwefantasy on 14/1/2018.
*/
class SQLFPGrowth extends SQLAlg with Functions {
override def train(df: DataFrame, path: String, params: Map[String, String]): DataFrame = {
val rfc = new FPGrowth()
configureModel(rfc, params)
val model = rfc.fit(df)
model.write.overwrite().save(path)
emptyDataFrame()(df)
}
override def load(sparkSession: SparkSession, path: String, params: Map[String, String]): Any = {
val model = FPGrowthModel.load(path)
model
}
override def predict(sparkSession: SparkSession, _model: Any, name: String, params: Map[String, String]): UserDefinedFunction = {
val model = _model.asInstanceOf[FPGrowthModel]
val rules: Array[(Seq[String], Seq[String])] = model.associationRules.select("antecedent", "consequent")
.rdd.map(r => (r.getSeq(0), r.getSeq(1)))
.collect().asInstanceOf[Array[(Seq[String], Seq[String])]]
val brRules = sparkSession.sparkContext.broadcast(rules)
val f = (items: Seq[String]) => {
if (items != null) {
val itemset = items.toSet
brRules.value.flatMap(rule =>
if (items != null && rule._1.forall(item => itemset.contains(item))) {
rule._2.filter(item => !itemset.contains(item))
} else {
Seq.empty
}).distinct
} else {
Seq.empty
}
}
UserDefinedFunction(f, ArrayType(StringType), Some(Seq(ArrayType(StringType))))
}
}