Connect R to Spark using Sparklyr (spark_home file.path error)?
I am a starter for big data. Now I’m trying to connect R to Spark using Sparklyr (1.8.6), and have the following error message I haven’t found answers in stackoverflow.
How to do parallel model training for each spark partition in using sparklyR for R?
To me it seems like it should be a straight-out-of-the-box feature when working with spark data tables, but I can simply not find out how to train one model per spark partition (or some sort of grouping) using sparklyR
in R.