Skip to content

Latest commit

 

History

History
55 lines (44 loc) · 1.04 KB

use-guide.md

File metadata and controls

55 lines (44 loc) · 1.04 KB

use-guide

  1. Setup spark cluster.
docker-compose up -d
  1. Enter spark-master.
docker exec -it {container-id} /bin/bash
  1. List all containers。
docker ps -a
  1. Stop container.
docker stop {container-id}
  1. Delete container.
docker rm {container-id}
  1. Spark management webstie.

Chrome: "http://127.0.0.1:8080"

  1. Test spark cluster
# login spark master
docker exec -it {container-id} /bin/bash
cd spark
bin/spark-submit --class org.apache.spark.examples.SparkPi --master spark://spark-master:7077 examples/jars/spark-examples_2.11-2.4.0.jar 100 2>&1 | grep "Pi is roughly"
  1. Spark python shell.
pyspark --master spark://spark-master:7077
#input python shell, compute pi
import random
NUM_SAMPLES = 1000000
def inside(p):
    x, y = random.random(), random.random()
    return x*x + y*y < 1

count = sc.parallelize(xrange(0, NUM_SAMPLES)) \
             .filter(inside).count()
print "Pi is roughly %f" % (4.0 * count / NUM_SAMPLES)
#output: Pi is roughly 3.145192