-
Try to run
$YOUR_SPARK_PATH/bin/spark-shell
$YOUR_SPARK_PATH/bin/pyspark
- Once it is loaded you will see scala console:
scala>
Type 'sc'
and press enter, you should see:
scala> sc
res0: org.apache.spark.SparkContext = org.apache.spark.SparkContext@1f60824e
-
Start spark locally.
1.1. Start the master
$YOUR_SPARK_PATH/sbin/start-master.sh
1.2. After couple of minutes the http://localhost:8080 will be available. It is a spark ui.
1.3. Start workers(as many as you like, 2 or 3):
$YOUR_SPARK_PATH/bin/spark-class org.apache.spark.deploy.worker.Worker $SPARK_MASTER_URL &
where $SPARK_MASTER_URL is in the [web ui](http://localhost:8080)
- Submit a test job
$YOUR_SPARK_PATH/bin/spark-submit --master $SPARK_MASTER_URL job.py $SPARK_MASTER_URL
Let's go through the official tutorial