0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012
0013
0014
0015
0016
0017
0018 package org.apache.spark.examples.ml;
0019
0020
0021 import java.util.Arrays;
0022 import java.util.List;
0023
0024 import org.apache.spark.ml.feature.HashingTF;
0025 import org.apache.spark.ml.feature.IDF;
0026 import org.apache.spark.ml.feature.IDFModel;
0027 import org.apache.spark.ml.feature.Tokenizer;
0028 import org.apache.spark.sql.Dataset;
0029 import org.apache.spark.sql.Row;
0030 import org.apache.spark.sql.RowFactory;
0031 import org.apache.spark.sql.SparkSession;
0032 import org.apache.spark.sql.types.DataTypes;
0033 import org.apache.spark.sql.types.Metadata;
0034 import org.apache.spark.sql.types.StructField;
0035 import org.apache.spark.sql.types.StructType;
0036
0037
0038 public class JavaTfIdfExample {
0039 public static void main(String[] args) {
0040 SparkSession spark = SparkSession
0041 .builder()
0042 .appName("JavaTfIdfExample")
0043 .getOrCreate();
0044
0045
0046 List<Row> data = Arrays.asList(
0047 RowFactory.create(0.0, "Hi I heard about Spark"),
0048 RowFactory.create(0.0, "I wish Java could use case classes"),
0049 RowFactory.create(1.0, "Logistic regression models are neat")
0050 );
0051 StructType schema = new StructType(new StructField[]{
0052 new StructField("label", DataTypes.DoubleType, false, Metadata.empty()),
0053 new StructField("sentence", DataTypes.StringType, false, Metadata.empty())
0054 });
0055 Dataset<Row> sentenceData = spark.createDataFrame(data, schema);
0056
0057 Tokenizer tokenizer = new Tokenizer().setInputCol("sentence").setOutputCol("words");
0058 Dataset<Row> wordsData = tokenizer.transform(sentenceData);
0059
0060 int numFeatures = 20;
0061 HashingTF hashingTF = new HashingTF()
0062 .setInputCol("words")
0063 .setOutputCol("rawFeatures")
0064 .setNumFeatures(numFeatures);
0065
0066 Dataset<Row> featurizedData = hashingTF.transform(wordsData);
0067
0068
0069 IDF idf = new IDF().setInputCol("rawFeatures").setOutputCol("features");
0070 IDFModel idfModel = idf.fit(featurizedData);
0071
0072 Dataset<Row> rescaledData = idfModel.transform(featurizedData);
0073 rescaledData.select("label", "features").show();
0074
0075
0076 spark.stop();
0077 }
0078 }