0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012
0013
0014
0015
0016
0017
0018 package org.apache.spark.ml.classification;
0019
0020 import java.util.HashMap;
0021 import java.util.Map;
0022
0023 import org.junit.Test;
0024
0025 import org.apache.spark.SharedSparkSession;
0026 import org.apache.spark.api.java.JavaRDD;
0027 import org.apache.spark.ml.feature.LabeledPoint;
0028 import org.apache.spark.ml.tree.impl.TreeTests;
0029 import org.apache.spark.sql.Dataset;
0030 import org.apache.spark.sql.Row;
0031
0032 public class JavaDecisionTreeClassifierSuite extends SharedSparkSession {
0033
0034 @Test
0035 public void runDT() {
0036 int nPoints = 20;
0037 double A = 2.0;
0038 double B = -1.5;
0039
0040 JavaRDD<LabeledPoint> data = jsc.parallelize(
0041 LogisticRegressionSuite.generateLogisticInputAsList(A, B, nPoints, 42), 2).cache();
0042 Map<Integer, Integer> categoricalFeatures = new HashMap<>();
0043 Dataset<Row> dataFrame = TreeTests.setMetadata(data, categoricalFeatures, 2);
0044
0045
0046 DecisionTreeClassifier dt = new DecisionTreeClassifier()
0047 .setMaxDepth(2)
0048 .setMaxBins(10)
0049 .setMinInstancesPerNode(5)
0050 .setMinInfoGain(0.0)
0051 .setMaxMemoryInMB(256)
0052 .setCacheNodeIds(false)
0053 .setCheckpointInterval(10)
0054 .setMaxDepth(2);
0055 for (String impurity : DecisionTreeClassifier.supportedImpurities()) {
0056 dt.setImpurity(impurity);
0057 }
0058 DecisionTreeClassificationModel model = dt.fit(dataFrame);
0059
0060 model.transform(dataFrame);
0061 model.numNodes();
0062 model.depth();
0063 model.toDebugString();
0064
0065
0066
0067
0068
0069
0070
0071
0072
0073
0074
0075
0076
0077
0078 }
0079 }