The SparkConf doesn't allow you to set arbitrary variables.  You can use
SparkContext's HadoopRDD and create a JobConf (with whatever variables you
want), and then grab them out of the JobConf in your RecordReader.

On Sun, Feb 22, 2015 at 4:28 PM, hnahak <harihar1...@gmail.com> wrote:

> Hi,
>
> I have written custom InputFormat and RecordReader for Spark, I need  to
> use
> user variables from spark client program.
>
> I added them in SparkConf
>
>  val sparkConf = new
> SparkConf().setAppName(args(0)).set("developer","MyName")
>
> *and in InputFormat class*
>
>         protected boolean isSplitable(JobContext context, Path filename) {
>
>
> System.out.println("######################################### Developer "
> + context.getConfiguration().get("developer") );
>                 return false;
>         }
>
> but its return me *null* , is there any way I can pass user variables to my
> custom code?
>
> Thanks !!
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/How-to-send-user-variables-from-Spark-client-to-custom-InputFormat-or-RecordReader-tp21755.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>

Reply via email to