« RStudio v1.0 released | Main | How did the election forecasts get it so wrong? »

November 08, 2016

Comments

Feed You can follow this conversation by subscribing to the comment feed for this post.

There is no "R Server on Spark" option in my Cluster Configuration - there is R Server OR Spark...

Also,

Error in force(code) :
Failed while connecting to sparklyr to port (8880) for sessionid (6826): Gateway in port (8880) did not respond.
Path: C:\Users\jvangeete\Apache\Spark-2.0.2\bin\spark-submit2.cmd
Parameters: --class, sparklyr.Backend, "C:\Users\jvangeete\Documents\R\win-library\3.3\sparklyr\java\sparklyr-2.0-2.11.jar", 8880, 6826


---- Output Log ----
The system cannot find the path specified.

---- Error Log ----

Thanks for the comment, Jeff!
You're right, the portal has changed slightly. The correct choice for the cluster configuration is just "R Server". It will automatically build on top of Spark (1.6.2 for a few more days, then 2.0!).
Did that error come when you tried to use `spark_connect("yarn-client")`? Could you try
`sc <- spark_connect("yarn",
config = list(default = list(spark.submit.deployMode= "client"))` instead? It appears you are using Spark 2.0 clusters rather than the 1.6.2 clusters I used for this post.

Thanks,
Ali

The comments to this entry are closed.

Search Revolutions Blog




Got comments or suggestions for the blog editor?
Email David Smith.
Follow revodavid on Twitter Follow David on Twitter: @revodavid
Get this blog via email with Blogtrottr