Interactive analysis with the SparkR shell
The entry point into SparkR is the SparkContext which connects the R program to a Spark Cluster. When working with the SparkR shell, SQLContext and SparkContext are already available. SparkR's shell provides a simple way to learn the API, as well as a powerful tool to analyze data interactively.
Getting ready
To step through this recipe, you will need a running Spark Cluster either in pseudo distributed mode or in one of the distributed modes, that is, standalone, YARN, or Mesos.
How to do it…
In this recipe, we’ll see how to start SparkR interactive shell using Spark 1.6.0:
- Start the SparkR shell by running the following in the SparkR package directory:
/bigdata/spark-1.6.0-bin-hadoop2.6$ ./bin/sparkR --master spark://192.168.0.118:7077 R version 3.2.3 (2015-12-10) -- "Wooden Christmas-Tree" Copyright (C) 2015 The R Foundation for Statistical Computing Platform: x86_64-pc-linux-gnu (64-bit) R is free...