that variable "x" would be a DataFrame which is an alias of Dataset in the
last versions. you can do your map operation by doing x.map(case
Row(f1:String, f2:Int, ) => [your code]). f1 and f2 stands for the
columns of your dataset with the type. in the code you can use f1 and f2 as
variables
You are using spark as a library but it is much more than that. The book
"learning Spark" is very well done and it helped me a lot starting with
spark. Maybe you should start from there.
Those are the issues in your code:
Basically, you generally don't execute spark code like that. You could
Thanks for the advice Diego, that was very helpful. How could I read the
csv as a dataset though? I need to do a map operation over the dataset, I
just coded up an example to illustrate the issue
On Mar 22, 2017 6:43 PM, "Diego Fanesi" wrote:
> You are using spark as a