|
||||
0001 # 0002 # Licensed to the Apache Software Foundation (ASF) under one or more 0003 # contributor license agreements. See the NOTICE file distributed with 0004 # this work for additional information regarding copyright ownership. 0005 # The ASF licenses this file to You under the Apache License, Version 2.0 0006 # (the "License"); you may not use this file except in compliance with 0007 # the License. You may obtain a copy of the License at 0008 # 0009 # http://www.apache.org/licenses/LICENSE-2.0 0010 # 0011 # Unless required by applicable law or agreed to in writing, software 0012 # distributed under the License is distributed on an "AS IS" BASIS, 0013 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 0014 # See the License for the specific language governing permissions and 0015 # limitations under the License. 0016 # 0017 0018 from __future__ import print_function 0019 0020 import sys 0021 from random import random 0022 from operator import add 0023 0024 from pyspark.sql import SparkSession 0025 0026 0027 if __name__ == "__main__": 0028 """ 0029 Usage: pi [partitions] 0030 """ 0031 spark = SparkSession\ 0032 .builder\ 0033 .appName("PythonPi")\ 0034 .getOrCreate() 0035 0036 partitions = int(sys.argv[1]) if len(sys.argv) > 1 else 2 0037 n = 100000 * partitions 0038 0039 def f(_): 0040 x = random() * 2 - 1 0041 y = random() * 2 - 1 0042 return 1 if x ** 2 + y ** 2 <= 1 else 0 0043 0044 count = spark.sparkContext.parallelize(range(1, n + 1), partitions).map(f).reduce(add) 0045 print("Pi is roughly %f" % (4.0 * count / n)) 0046 0047 spark.stop()
[ Source navigation ] | [ Diff markup ] | [ Identifier search ] | [ general search ] |
This page was automatically generated by the 2.1.0 LXR engine. The LXR team |