0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012
0013
0014
0015
0016
0017
0018 package org.apache.spark.examples.ml;
0019
0020
0021 import java.util.Arrays;
0022 import java.util.List;
0023
0024 import org.apache.spark.ml.feature.Imputer;
0025 import org.apache.spark.ml.feature.ImputerModel;
0026 import org.apache.spark.sql.Dataset;
0027 import org.apache.spark.sql.Row;
0028 import org.apache.spark.sql.RowFactory;
0029 import org.apache.spark.sql.SparkSession;
0030 import org.apache.spark.sql.types.*;
0031
0032
0033 import static org.apache.spark.sql.types.DataTypes.*;
0034
0035
0036
0037
0038
0039
0040 public class JavaImputerExample {
0041 public static void main(String[] args) {
0042 SparkSession spark = SparkSession
0043 .builder()
0044 .appName("JavaImputerExample")
0045 .getOrCreate();
0046
0047
0048 List<Row> data = Arrays.asList(
0049 RowFactory.create(1.0, Double.NaN),
0050 RowFactory.create(2.0, Double.NaN),
0051 RowFactory.create(Double.NaN, 3.0),
0052 RowFactory.create(4.0, 4.0),
0053 RowFactory.create(5.0, 5.0)
0054 );
0055 StructType schema = new StructType(new StructField[]{
0056 createStructField("a", DoubleType, false),
0057 createStructField("b", DoubleType, false)
0058 });
0059 Dataset<Row> df = spark.createDataFrame(data, schema);
0060
0061 Imputer imputer = new Imputer()
0062 .setInputCols(new String[]{"a", "b"})
0063 .setOutputCols(new String[]{"out_a", "out_b"});
0064
0065 ImputerModel model = imputer.fit(df);
0066 model.transform(df).show();
0067
0068
0069 spark.stop();
0070 }
0071 }