Unable to launch SparkR in RStudio

前端 未结 7 983
一向
一向 2021-02-19 07:52

After long and difficult installation process of SparkR i getting into new problems of launching SparkR.

My Settings

R 3.2.0    
RStudio 0.98.1103    
Rt         


        
相关标签:
7条回答
  • 2021-02-19 08:17

    I had a similar issue. In my case the problem was with the hyphen ('-').
    by changing the code :

    sc <- sparkR.init(master = "local[*]",sparkPackages = c("com.databricks:spark-csv_2.11-1.4.0"))
    

    to:

    sc <- sparkR.init(master = "local[*]",sparkPackages = c("com.databricks:spark-csv_2.11:1.4.0"))
    

    worked for me. Do you notice the change?

    P.S.: Do copy the jar in your SPARK_HOME\lib folder

    Edit 1: Also, check that you have configured your "HADOOP_HOME"


    Hope this helps.

    0 讨论(0)
  • 2021-02-19 08:18

    I had the same issue and my spark-submit.cmd file was also not executing from the command line. Following steps worked for me

    Go to your environment variables and in the system variables select variable name PATH. Along with other values add c:/Windows/System32/ separated by a semicolon. This made my spark-submit.cmd run from command line and eventually from the Rstudio.

    I have realized that we get the above issue only if all the required path values are not specified. Ensure all your path values(R, Rtools) are specified in the environment variables. For instance my Rtools path was c:\Rtools\bin;c:\Rtools\gcc-4.6.3\bin

    I hope this helps.

    0 讨论(0)
  • 2021-02-19 08:20

    That didn't work for me. If anyone has the same problem, try to give execute permissions to c:/sparkpath/bin/spark-submit.cmd.

    0 讨论(0)
  • 2021-02-19 08:22

    I think it was a bug that has now been resolved. Try the following,

    Sys.setenv(SPARK_HOME="C:\\spark-1.4.0")
    
    .libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"), .libPaths()))
    
    library("SparkR", lib.loc="C:\\spark-1.4.0\\lib") # The use of \\ is for windows environment.
    
    library(SparkR)
    
    sc=sparkR.init(master="local")
    

    Launching java with spark-submit command C:\spark-1.4.0/bin/spark-submit.cmd sparkr-shell

    C:\Users\Ashish\AppData\Local\Temp\RtmpWqFsOB\backend_portbdc329477c6

    Hope this helps.

    0 讨论(0)
  • 2021-02-19 08:43

    I had exact same issue. I can start SparkR in command line, but not in RStudio in Windows. And here is the solution works for me.

    1. clean up all the paths you set when you tried to fix this issue. This including the paths you set in the windows environment from window control panel and uses Sys.unsetenv() to unset the SPARK_HOME.

    2. find out your RStudio default working directory by using getwd() in RStudio. And then create a .Rprofile file in this directory. Put the following line in this file: .libPaths("C:/Apache/Spark-1.5.1/R/lib")

    3. In window control panel->System->Advanced system settings->Environment Variables, add this ";C:\Apache\Spark-1.5.1\bin" at the end of your exsiting PATH variable.

    4. Start RStudio, if you type .libPaths(), you can see the SparkR library path is already in the library path

    5. use library(SparkR) to load SparkR library

    6. sc=sparkR.init(master="local")

    I tried this on both Spark 1.4.1 and 1.5.1, they both work fine. I hope this can help whoever still having issue after all the suggestion above.

    0 讨论(0)
  • 2021-02-19 08:43

    The following solution will work for Mac OS.

    After installing Hadoop followed by Spark.

    spark_path <- strsplit(system("brew info apache-spark",intern=T)[4],' ')[[1]][1] # Get your spark path .libPaths(c(file.path(spark_path,"libexec", "R", "lib"), .libPaths())) library(SparkR

    0 讨论(0)
提交回复
热议问题