Re: Re: RE: Error not found value sqlContext

2015-11-23 Thread satish chandra j
Thanks for all the support.
It was a code issue which I overlooked it

Regards,
Satish Chandra

On Mon, Nov 23, 2015 at 3:49 PM, satish chandra j 
wrote:

> Sorry, just to understand my issue.if Eclipse could not understand
> Scala syntax properly than it should error for the other Spark Job which is
> fetching data from a RDBM and printing the output in console. I think there
> is some dependencies which are missing due to which "import
> sqlContext.implicits._" is not recognized during compile time
>
> Please let me know if any further inputs needed to fix the same
>
> Regards,
> Satish Chandra
>
> On Mon, Nov 23, 2015 at 3:29 PM, prosp4300  wrote:
>
>>
>>
>> So it is actually a compile time error in Eclipse, instead of jar
>> generation from Eclipse, you can try to use sbt to assembly your jar, looks
>> like your Eclipse does not recognize the Scala syntax properly.
>>
>>
>>
>> At 2015-11-20 21:36:55, "satish chandra j" 
>> wrote:
>>
>> HI All,
>> I am getting this error while generating executable Jar file itself in
>> Eclipse, if the Spark Application code has "import sqlContext.implicits._"
>> line in it. Spark Applicaiton code  works fine if the above mentioned line
>> does not exist as I have tested by fetching data from an RDBMS by
>> implementing JDBCRDD
>>
>> I tried couple of DataFrame related methods for which most of them errors
>> stating that method has been overloaded
>>
>> Please let me know if any further inputs needed to analyze it
>>
>> Regards,
>> Satish Chandra
>>
>> On Fri, Nov 20, 2015 at 5:46 PM, prosp4300  wrote:
>>
>>>
>>> Looks like a classpath problem, if you can provide the command you used
>>> to run your application and environment variable SPARK_HOME, it will help
>>> others to identify the root problem
>>>
>>>
>>> 在2015年11月20日 18:59,Satish  写道:
>>>
>>> Hi Michael,
>>> As my current Spark version is 1.4.0 than why it error out as "error:
>>> not found: value sqlContext" when I have "import sqlContext.implicits._" in
>>> my Spark Job
>>>
>>> Regards
>>> Satish Chandra
>>> --
>>> From: Michael Armbrust 
>>> Sent: ‎20-‎11-‎2015 01:36
>>> To: satish chandra j 
>>> Cc: user ; hari krishna 
>>> Subject: Re: Error not found value sqlContext
>>>
>>>
>>> http://spark.apache.org/docs/latest/sql-programming-guide.html#upgrading-from-spark-sql-10-12-to-13
>>>
>>> On Thu, Nov 19, 2015 at 4:19 AM, satish chandra j <
>>> jsatishchan...@gmail.com> wrote:
>>>
>>>> HI All,
>>>> we have recently migrated from Spark 1.2.1 to Spark 1.4.0, I am
>>>> fetching data from an RDBMS using JDBCRDD and register it as temp table to
>>>> perform SQL query
>>>>
>>>> Below approach is working fine in Spark 1.2.1:
>>>>
>>>> JDBCRDD --> apply map using Case Class --> apply createSchemaRDD -->
>>>> registerTempTable --> perform SQL Query
>>>>
>>>> but now as createSchemaRDD is not supported in Spark 1.4.0
>>>>
>>>> JDBCRDD --> apply map using Case Class with* .toDF()* -->
>>>> registerTempTable --> perform SQL query on temptable
>>>>
>>>>
>>>> JDBCRDD --> apply map using Case Class --> RDD*.toDF()*.registerTempTable
>>>> --> perform SQL query on temptable
>>>>
>>>> Only solution I get everywhere is to  use "import
>>>> sqlContext.implicits._" after val SQLContext = new
>>>> org.apache.spark.sql.SQLContext(sc)
>>>>
>>>> But it errors with the two generic errors
>>>>
>>>> *1. error: not found: value sqlContext*
>>>>
>>>> *2. value toDF is not a member of org.apache.spark.rdd.RDD*
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>>
>>
>>
>>
>>
>
>


Re: Re: RE: Error not found value sqlContext

2015-11-23 Thread satish chandra j
Sorry, just to understand my issue.if Eclipse could not understand
Scala syntax properly than it should error for the other Spark Job which is
fetching data from a RDBM and printing the output in console. I think there
is some dependencies which are missing due to which "import
sqlContext.implicits._" is not recognized during compile time

Please let me know if any further inputs needed to fix the same

Regards,
Satish Chandra

On Mon, Nov 23, 2015 at 3:29 PM, prosp4300  wrote:

>
>
> So it is actually a compile time error in Eclipse, instead of jar
> generation from Eclipse, you can try to use sbt to assembly your jar, looks
> like your Eclipse does not recognize the Scala syntax properly.
>
>
>
> At 2015-11-20 21:36:55, "satish chandra j" 
> wrote:
>
> HI All,
> I am getting this error while generating executable Jar file itself in
> Eclipse, if the Spark Application code has "import sqlContext.implicits._"
> line in it. Spark Applicaiton code  works fine if the above mentioned line
> does not exist as I have tested by fetching data from an RDBMS by
> implementing JDBCRDD
>
> I tried couple of DataFrame related methods for which most of them errors
> stating that method has been overloaded
>
> Please let me know if any further inputs needed to analyze it
>
> Regards,
> Satish Chandra
>
> On Fri, Nov 20, 2015 at 5:46 PM, prosp4300  wrote:
>
>>
>> Looks like a classpath problem, if you can provide the command you used
>> to run your application and environment variable SPARK_HOME, it will help
>> others to identify the root problem
>>
>>
>> 在2015年11月20日 18:59,Satish  写道:
>>
>> Hi Michael,
>> As my current Spark version is 1.4.0 than why it error out as "error: not
>> found: value sqlContext" when I have "import sqlContext.implicits._" in my
>> Spark Job
>>
>> Regards
>> Satish Chandra
>> --
>> From: Michael Armbrust 
>> Sent: ‎20-‎11-‎2015 01:36
>> To: satish chandra j 
>> Cc: user ; hari krishna 
>> Subject: Re: Error not found value sqlContext
>>
>>
>> http://spark.apache.org/docs/latest/sql-programming-guide.html#upgrading-from-spark-sql-10-12-to-13
>>
>> On Thu, Nov 19, 2015 at 4:19 AM, satish chandra j <
>> jsatishchan...@gmail.com> wrote:
>>
>>> HI All,
>>> we have recently migrated from Spark 1.2.1 to Spark 1.4.0, I am fetching
>>> data from an RDBMS using JDBCRDD and register it as temp table to perform
>>> SQL query
>>>
>>> Below approach is working fine in Spark 1.2.1:
>>>
>>> JDBCRDD --> apply map using Case Class --> apply createSchemaRDD -->
>>> registerTempTable --> perform SQL Query
>>>
>>> but now as createSchemaRDD is not supported in Spark 1.4.0
>>>
>>> JDBCRDD --> apply map using Case Class with* .toDF()* -->
>>> registerTempTable --> perform SQL query on temptable
>>>
>>>
>>> JDBCRDD --> apply map using Case Class --> RDD*.toDF()*.registerTempTable
>>> --> perform SQL query on temptable
>>>
>>> Only solution I get everywhere is to  use "import
>>> sqlContext.implicits._" after val SQLContext = new
>>> org.apache.spark.sql.SQLContext(sc)
>>>
>>> But it errors with the two generic errors
>>>
>>> *1. error: not found: value sqlContext*
>>>
>>> *2. value toDF is not a member of org.apache.spark.rdd.RDD*
>>>
>>>
>>>
>>>
>>>
>>>
>>
>>
>>
>
>
>
>


Re:Re: RE: Error not found value sqlContext

2015-11-23 Thread prosp4300


So it is actually a compile time error in Eclipse, instead of jar generation 
from Eclipse, you can try to use sbt to assembly your jar, looks like your 
Eclipse does not recognize the Scala syntax properly.



At 2015-11-20 21:36:55, "satish chandra j"  wrote:

HI All,
I am getting this error while generating executable Jar file itself in Eclipse, 
if the Spark Application code has "import sqlContext.implicits._" line in it. 
Spark Applicaiton code  works fine if the above mentioned line does not exist 
as I have tested by fetching data from an RDBMS by implementing JDBCRDD


I tried couple of DataFrame related methods for which most of them errors 
stating that method has been overloaded


Please let me know if any further inputs needed to analyze it


Regards,
Satish Chandra


On Fri, Nov 20, 2015 at 5:46 PM, prosp4300  wrote:


Looks like a classpath problem, if you can provide the command you used to run 
your application and environment variable SPARK_HOME, it will help others to 
identify the root problem




在2015年11月20日 18:59,Satish 写道:
Hi Michael,
As my current Spark version is 1.4.0 than why it error out as "error: not 
found: value sqlContext" when I have "import sqlContext.implicits._" in my 
Spark Job

Regards
Satish Chandra
From: Michael Armbrust
Sent: ‎20-‎11-‎2015 01:36
To: satish chandra j
Cc: user; hari krishna
Subject: Re: Error not found value sqlContext


http://spark.apache.org/docs/latest/sql-programming-guide.html#upgrading-from-spark-sql-10-12-to-13



On Thu, Nov 19, 2015 at 4:19 AM, satish chandra j  
wrote:

HI All,
we have recently migrated from Spark 1.2.1 to Spark 1.4.0, I am fetching data 
from an RDBMS using JDBCRDD and register it as temp table to perform SQL query


Below approach is working fine in Spark 1.2.1:


JDBCRDD --> apply map using Case Class --> apply createSchemaRDD --> 
registerTempTable --> perform SQL Query


but now as createSchemaRDD is not supported in Spark 1.4.0



JDBCRDD --> apply map using Case Class with .toDF() --> registerTempTable --> 
perform SQL query on temptable




JDBCRDD --> apply map using Case Class --> RDD.toDF().registerTempTable --> 
perform SQL query on temptable



Only solution I get everywhere is to  use "import sqlContext.implicits._" after 
val SQLContext = new org.apache.spark.sql.SQLContext(sc)


But it errors with the two generic errors


1. error: not found: value sqlContext


2. value toDF is not a member of org.apache.spark.rdd.RDD


















Re: RE: Error not found value sqlContext

2015-11-20 Thread satish chandra j
HI All,
I am getting this error while generating executable Jar file itself in
Eclipse, if the Spark Application code has "import sqlContext.implicits._"
line in it. Spark Applicaiton code  works fine if the above mentioned line
does not exist as I have tested by fetching data from an RDBMS by
implementing JDBCRDD

I tried couple of DataFrame related methods for which most of them errors
stating that method has been overloaded

Please let me know if any further inputs needed to analyze it

Regards,
Satish Chandra

On Fri, Nov 20, 2015 at 5:46 PM, prosp4300  wrote:

>
> Looks like a classpath problem, if you can provide the command you used to
> run your application and environment variable SPARK_HOME, it will help
> others to identify the root problem
>
>
> 在2015年11月20日 18:59,Satish  写道:
>
> Hi Michael,
> As my current Spark version is 1.4.0 than why it error out as "error: not
> found: value sqlContext" when I have "import sqlContext.implicits._" in my
> Spark Job
>
> Regards
> Satish Chandra
> --
> From: Michael Armbrust 
> Sent: ‎20-‎11-‎2015 01:36
> To: satish chandra j 
> Cc: user ; hari krishna 
> Subject: Re: Error not found value sqlContext
>
>
> http://spark.apache.org/docs/latest/sql-programming-guide.html#upgrading-from-spark-sql-10-12-to-13
>
> On Thu, Nov 19, 2015 at 4:19 AM, satish chandra j <
> jsatishchan...@gmail.com> wrote:
>
>> HI All,
>> we have recently migrated from Spark 1.2.1 to Spark 1.4.0, I am fetching
>> data from an RDBMS using JDBCRDD and register it as temp table to perform
>> SQL query
>>
>> Below approach is working fine in Spark 1.2.1:
>>
>> JDBCRDD --> apply map using Case Class --> apply createSchemaRDD -->
>> registerTempTable --> perform SQL Query
>>
>> but now as createSchemaRDD is not supported in Spark 1.4.0
>>
>> JDBCRDD --> apply map using Case Class with* .toDF()* -->
>> registerTempTable --> perform SQL query on temptable
>>
>>
>> JDBCRDD --> apply map using Case Class --> RDD*.toDF()*.registerTempTable
>> --> perform SQL query on temptable
>>
>> Only solution I get everywhere is to  use "import sqlContext.implicits._"
>> after val SQLContext = new org.apache.spark.sql.SQLContext(sc)
>>
>> But it errors with the two generic errors
>>
>> *1. error: not found: value sqlContext*
>>
>> *2. value toDF is not a member of org.apache.spark.rdd.RDD*
>>
>>
>>
>>
>>
>>
>
>
>


回复:RE: Error not found value sqlContext

2015-11-20 Thread prosp4300

Looks like a classpath problem, if you can provide the command you used to run 
your application and environment variable SPARK_HOME, it will help others to 
identify the root problem



在2015年11月20日 18:59,Satish 写道:
Hi Michael,
As my current Spark version is 1.4.0 than why it error out as "error: not 
found: value sqlContext" when I have "import sqlContext.implicits._" in my 
Spark Job

Regards
Satish Chandra
From: Michael Armbrust
Sent: ‎20-‎11-‎2015 01:36
To: satish chandra j
Cc: user; hari krishna
Subject: Re: Error not found value sqlContext


http://spark.apache.org/docs/latest/sql-programming-guide.html#upgrading-from-spark-sql-10-12-to-13



On Thu, Nov 19, 2015 at 4:19 AM, satish chandra j  
wrote:

HI All,
we have recently migrated from Spark 1.2.1 to Spark 1.4.0, I am fetching data 
from an RDBMS using JDBCRDD and register it as temp table to perform SQL query


Below approach is working fine in Spark 1.2.1:


JDBCRDD --> apply map using Case Class --> apply createSchemaRDD --> 
registerTempTable --> perform SQL Query


but now as createSchemaRDD is not supported in Spark 1.4.0



JDBCRDD --> apply map using Case Class with .toDF() --> registerTempTable --> 
perform SQL query on temptable




JDBCRDD --> apply map using Case Class --> RDD.toDF().registerTempTable --> 
perform SQL query on temptable



Only solution I get everywhere is to  use "import sqlContext.implicits._" after 
val SQLContext = new org.apache.spark.sql.SQLContext(sc)


But it errors with the two generic errors


1. error: not found: value sqlContext


2. value toDF is not a member of org.apache.spark.rdd.RDD













RE: Error not found value sqlContext

2015-11-20 Thread Satish
Hi Michael,
As my current Spark version is 1.4.0 than why it error out as "error: not 
found: value sqlContext" when I have "import sqlContext.implicits._" in my 
Spark Job

Regards 
Satish Chandra

-Original Message-
From: "Michael Armbrust" 
Sent: ‎20-‎11-‎2015 01:36
To: "satish chandra j" 
Cc: "user" ; "hari krishna" 
Subject: Re: Error not found value sqlContext

http://spark.apache.org/docs/latest/sql-programming-guide.html#upgrading-from-spark-sql-10-12-to-13



On Thu, Nov 19, 2015 at 4:19 AM, satish chandra j  
wrote:

HI All,
we have recently migrated from Spark 1.2.1 to Spark 1.4.0, I am fetching data 
from an RDBMS using JDBCRDD and register it as temp table to perform SQL query


Below approach is working fine in Spark 1.2.1:


JDBCRDD --> apply map using Case Class --> apply createSchemaRDD --> 
registerTempTable --> perform SQL Query


but now as createSchemaRDD is not supported in Spark 1.4.0



JDBCRDD --> apply map using Case Class with .toDF() --> registerTempTable --> 
perform SQL query on temptable




JDBCRDD --> apply map using Case Class --> RDD.toDF().registerTempTable --> 
perform SQL query on temptable



Only solution I get everywhere is to  use "import sqlContext.implicits._" after 
val SQLContext = new org.apache.spark.sql.SQLContext(sc)


But it errors with the two generic errors


1. error: not found: value sqlContext


2. value toDF is not a member of org.apache.spark.rdd.RDD

Re: Error not found value sqlContext

2015-11-19 Thread Michael Armbrust
http://spark.apache.org/docs/latest/sql-programming-guide.html#upgrading-from-spark-sql-10-12-to-13

On Thu, Nov 19, 2015 at 4:19 AM, satish chandra j 
wrote:

> HI All,
> we have recently migrated from Spark 1.2.1 to Spark 1.4.0, I am fetching
> data from an RDBMS using JDBCRDD and register it as temp table to perform
> SQL query
>
> Below approach is working fine in Spark 1.2.1:
>
> JDBCRDD --> apply map using Case Class --> apply createSchemaRDD -->
> registerTempTable --> perform SQL Query
>
> but now as createSchemaRDD is not supported in Spark 1.4.0
>
> JDBCRDD --> apply map using Case Class with* .toDF()* -->
> registerTempTable --> perform SQL query on temptable
>
>
> JDBCRDD --> apply map using Case Class --> RDD*.toDF()*.registerTempTable
> --> perform SQL query on temptable
>
> Only solution I get everywhere is to  use "import sqlContext.implicits._"
> after val SQLContext = new org.apache.spark.sql.SQLContext(sc)
>
> But it errors with the two generic errors
>
> *1. error: not found: value sqlContext*
>
> *2. value toDF is not a member of org.apache.spark.rdd.RDD*
>
>
>
>
>
>


Error not found value sqlContext

2015-11-19 Thread satish chandra j
HI All,
we have recently migrated from Spark 1.2.1 to Spark 1.4.0, I am fetching
data from an RDBMS using JDBCRDD and register it as temp table to perform
SQL query

Below approach is working fine in Spark 1.2.1:

JDBCRDD --> apply map using Case Class --> apply createSchemaRDD -->
registerTempTable --> perform SQL Query

but now as createSchemaRDD is not supported in Spark 1.4.0

JDBCRDD --> apply map using Case Class with* .toDF()* --> registerTempTable
--> perform SQL query on temptable


JDBCRDD --> apply map using Case Class --> RDD*.toDF()*.registerTempTable
--> perform SQL query on temptable

Only solution I get everywhere is to  use "import sqlContext.implicits._"
after val SQLContext = new org.apache.spark.sql.SQLContext(sc)

But it errors with the two generic errors

*1. error: not found: value sqlContext*

*2. value toDF is not a member of org.apache.spark.rdd.RDD*