0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012
0013
0014
0015
0016
0017
0018 package org.apache.spark.examples.mllib;
0019
0020
0021 import java.util.Arrays;
0022
0023
0024 import org.apache.spark.SparkConf;
0025 import org.apache.spark.api.java.JavaSparkContext;
0026
0027 import org.apache.spark.api.java.JavaRDD;
0028 import org.apache.spark.mllib.feature.ElementwiseProduct;
0029 import org.apache.spark.mllib.linalg.Vector;
0030 import org.apache.spark.mllib.linalg.Vectors;
0031
0032
0033 public class JavaElementwiseProductExample {
0034 public static void main(String[] args) {
0035
0036 SparkConf conf = new SparkConf().setAppName("JavaElementwiseProductExample");
0037 JavaSparkContext jsc = new JavaSparkContext(conf);
0038
0039
0040
0041 JavaRDD<Vector> data = jsc.parallelize(Arrays.asList(
0042 Vectors.dense(1.0, 2.0, 3.0), Vectors.dense(4.0, 5.0, 6.0)));
0043 Vector transformingVector = Vectors.dense(0.0, 1.0, 2.0);
0044 ElementwiseProduct transformer = new ElementwiseProduct(transformingVector);
0045
0046
0047 JavaRDD<Vector> transformedData = transformer.transform(data);
0048 JavaRDD<Vector> transformedData2 = data.map(transformer::transform);
0049
0050
0051 System.out.println("transformedData: ");
0052 transformedData.foreach(System.out::println);
0053
0054 System.out.println("transformedData2: ");
0055 transformedData2.foreach(System.out::println);
0056
0057 jsc.stop();
0058 }
0059 }