0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012
0013
0014
0015
0016
0017
0018 package org.apache.spark.ml.classification;
0019
0020 import java.util.HashMap;
0021 import java.util.Map;
0022
0023 import org.junit.Test;
0024
0025 import org.apache.spark.SharedSparkSession;
0026 import org.apache.spark.api.java.JavaRDD;
0027 import org.apache.spark.ml.feature.LabeledPoint;
0028 import org.apache.spark.ml.tree.impl.TreeTests;
0029 import org.apache.spark.sql.Dataset;
0030 import org.apache.spark.sql.Row;
0031
0032 public class JavaGBTClassifierSuite extends SharedSparkSession {
0033
0034 @Test
0035 public void runDT() {
0036 int nPoints = 20;
0037 double A = 2.0;
0038 double B = -1.5;
0039
0040 JavaRDD<LabeledPoint> data = jsc.parallelize(
0041 LogisticRegressionSuite.generateLogisticInputAsList(A, B, nPoints, 42), 2).cache();
0042 Map<Integer, Integer> categoricalFeatures = new HashMap<>();
0043 Dataset<Row> dataFrame = TreeTests.setMetadata(data, categoricalFeatures, 2);
0044
0045
0046 GBTClassifier rf = new GBTClassifier()
0047 .setMaxDepth(2)
0048 .setMaxBins(10)
0049 .setMinInstancesPerNode(5)
0050 .setMinInfoGain(0.0)
0051 .setMaxMemoryInMB(256)
0052 .setCacheNodeIds(false)
0053 .setCheckpointInterval(10)
0054 .setSubsamplingRate(1.0)
0055 .setSeed(1234)
0056 .setMaxIter(3)
0057 .setStepSize(0.1)
0058 .setMaxDepth(2);
0059 for (String lossType : GBTClassifier.supportedLossTypes()) {
0060 rf.setLossType(lossType);
0061 }
0062 GBTClassificationModel model = rf.fit(dataFrame);
0063
0064 model.transform(dataFrame);
0065 model.totalNumNodes();
0066 model.toDebugString();
0067 model.trees();
0068 model.treeWeights();
0069
0070
0071
0072
0073
0074
0075
0076
0077
0078
0079
0080
0081
0082 }
0083 }