Re: Spark build 1.6.2 error

2016-09-03 Thread Nachiketa
I think the difference was the -Dscala2.11 to the command line.

I have seen this show up when I miss that.

Regards,
Nachiketa

On Sat 3 Sep, 2016, 12:14 PM Diwakar Dhanuskodi, <
diwakar.dhanusk...@gmail.com> wrote:

> Hi,
>
> Just re-ran again without killing zinc server process
>
> /make-distribution.sh --name custom-spark --tgz  -Phadoop-2.6 -Phive
> -Pyarn -Dmaven.version=3.0.4 -Dscala-2.11 -X -rf :spark-sql_2.11
>
> Build is success. Not sure how it worked with just re-running command
> again.
>
> On Sat, Sep 3, 2016 at 11:44 AM, Diwakar Dhanuskodi <
> diwakar.dhanusk...@gmail.com> wrote:
>
>> Hi,
>>
>> java version 7
>>
>> mvn command
>> ./make-distribution.sh --name custom-spark --tgz  -Phadoop-2.6 -Phive
>> -Phive-thriftserver -Pyarn -Dmaven.version=3.0.4
>>
>>
>> yes, I executed script to change scala version to 2.11
>> killed  "com.typesafe zinc.Nailgun" process
>>
>> re-ran mvn with below command again
>>
>> ./make-distribution.sh --name custom-spark --tgz  -Phadoop-2.6 -Phive
>> -Phive-thriftserver -Pyarn -Dmaven.version=3.0.4 -X -rf :spark-sql_2.11
>>
>> Getting same error
>>
>> [warn]
>> /home/cloudera/Downloads/spark-1.6.2/sql/core/src/main/scala/org/apache/spark/sql/sources/interfaces.scala:911:
>> method isDir in class FileStatus is deprecated: see corresponding Javadoc
>> for more information.
>> [warn] status.isDir,
>> [warn]^
>> [error] missing or invalid dependency detected while loading class file
>> 'WebUI.class'.
>> [error] Could not access term eclipse in package org,
>> [error] because it (or its dependencies) are missing. Check your build
>> definition for
>> [error] missing or conflicting dependencies. (Re-run with
>> `-Ylog-classpath` to see the problematic classpath.)
>> [error] A full rebuild may help if 'WebUI.class' was compiled against an
>> incompatible version of org.
>> [error] missing or invalid dependency detected while loading class file
>> 'WebUI.class'.
>> [error] Could not access term jetty in value org.eclipse,
>> [error] because it (or its dependencies) are missing. Check your build
>> definition for
>> [error] missing or conflicting dependencies. (Re-run with
>> `-Ylog-classpath` to see the problematic classpath.)
>> [error] A full rebuild may help if 'WebUI.class' was compiled against an
>> incompatible version of org.eclipse.
>> [warn] 17 warnings found
>> [error] two errors found
>> [debug] Compilation failed (CompilerInterface)
>> [error] Compile failed at Sep 3, 2016 11:28:34 AM [21.611s]
>> [INFO]
>> 
>> [INFO] Reactor Summary:
>> [INFO]
>> [INFO] Spark Project Parent POM .. SUCCESS
>> [5.583s]
>> [INFO] Spark Project Test Tags ... SUCCESS
>> [4.189s]
>> [INFO] Spark Project Launcher  SUCCESS
>> [12.226s]
>> [INFO] Spark Project Networking .. SUCCESS
>> [13.386s]
>> [INFO] Spark Project Shuffle Streaming Service ... SUCCESS
>> [6.723s]
>> [INFO] Spark Project Unsafe .. SUCCESS
>> [21.231s]
>> [INFO] Spark Project Core  SUCCESS
>> [3:46.334s]
>> [INFO] Spark Project Bagel ... SUCCESS
>> [7.032s]
>> [INFO] Spark Project GraphX .. SUCCESS
>> [19.558s]
>> [INFO] Spark Project Streaming ... SUCCESS
>> [50.452s]
>> [INFO] Spark Project Catalyst  SUCCESS
>> [1:14.172s]
>> [INFO] Spark Project SQL . FAILURE
>> [23.222s]
>> [INFO] Spark Project ML Library .. SKIPPED
>> [INFO] Spark Project Tools ... SKIPPED
>> [INFO] Spark Project Hive  SKIPPED
>> [INFO] Spark Project Docker Integration Tests  SKIPPED
>> [INFO] Spark Project REPL  SKIPPED
>> [INFO] Spark Project YARN Shuffle Service  SKIPPED
>> [INFO] Spark Project YARN  SKIPPED
>> [INFO] Spark Project Assembly  SKIPPED
>> [INFO] Spark Project External Twitter  SKIPPED
>> [INFO] Spark Project External Flume Sink . SKIPPED
>> [INFO] Spark Project External Flume .

Re: Spark build 1.6.2 error

2016-08-31 Thread Nachiketa
Hi Diwakar,

Could you please share the entire maven command that you are using to build
? And also the JDK version you are using ?

Also could you please confirm that you did execute the script for change
scala version to 2.11 before starting the build ? Thanks.

Regards,
Nachiketa

On Wed, Aug 31, 2016 at 2:00 AM, Diwakar Dhanuskodi <
diwakar.dhanusk...@gmail.com> wrote:

> Hi,
>
> While building Spark 1.6.2 , getting below error in spark-sql. Much
> appreciate for any help.
>
> ERROR] missing or invalid dependency detected while loading class file
> 'WebUI.class'.
> Could not access term eclipse in package org,
> because it (or its dependencies) are missing. Check your build definition
> for
> missing or conflicting dependencies. (Re-run with `-Ylog-classpath` to see
> the problematic classpath.)
> A full rebuild may help if 'WebUI.class' was compiled against an
> incompatible version of org.
> [ERROR] missing or invalid dependency detected while loading class file
> 'WebUI.class'.
> Could not access term jetty in value org.eclipse,
> because it (or its dependencies) are missing. Check your build definition
> for
> missing or conflicting dependencies. (Re-run with `-Ylog-classpath` to see
> the problematic classpath.)
> A full rebuild may help if 'WebUI.class' was compiled against an
> incompatible version of org.eclipse.
> [WARNING] 17 warnings found
> [ERROR] two errors found
> [INFO] 
> 
> [INFO] Reactor Summary:
> [INFO]
> [INFO] Spark Project Parent POM .. SUCCESS [4.399s]
> [INFO] Spark Project Test Tags ... SUCCESS [3.443s]
> [INFO] Spark Project Launcher  SUCCESS
> [10.131s]
> [INFO] Spark Project Networking .. SUCCESS
> [11.849s]
> [INFO] Spark Project Shuffle Streaming Service ... SUCCESS [6.641s]
> [INFO] Spark Project Unsafe .. SUCCESS
> [19.765s]
> [INFO] Spark Project Core  SUCCESS
> [4:16.511s]
> [INFO] Spark Project Bagel ... SUCCESS
> [13.401s]
> [INFO] Spark Project GraphX .. SUCCESS
> [1:08.824s]
> [INFO] Spark Project Streaming ... SUCCESS
> [2:18.844s]
> [INFO] Spark Project Catalyst  SUCCESS
> [2:43.695s]
> [INFO] Spark Project SQL . FAILURE
> [1:01.762s]
> [INFO] Spark Project ML Library .. SKIPPED
> [INFO] Spark Project Tools ... SKIPPED
> [INFO] Spark Project Hive  SKIPPED
> [INFO] Spark Project Docker Integration Tests  SKIPPED
> [INFO] Spark Project REPL  SKIPPED
> [INFO] Spark Project YARN Shuffle Service  SKIPPED
> [INFO] Spark Project YARN  SKIPPED
> [INFO] Spark Project Assembly  SKIPPED
> [INFO] Spark Project External Twitter  SKIPPED
> [INFO] Spark Project External Flume Sink . SKIPPED
> [INFO] Spark Project External Flume .. SKIPPED
> [INFO] Spark Project External Flume Assembly . SKIPPED
> [INFO] Spark Project External MQTT ... SKIPPED
> [INFO] Spark Project External MQTT Assembly .. SKIPPED
> [INFO] Spark Project External ZeroMQ . SKIPPED
> [INFO] Spark Project External Kafka .. SKIPPED
> [INFO] Spark Project Examples  SKIPPED
> [INFO] Spark Project External Kafka Assembly . SKIPPED
> [INFO] 
> 
> [INFO] BUILD FAILURE
> [INFO] 
> 
> [INFO] Total time: 12:40.525s
> [INFO] Finished at: Wed Aug 31 01:56:50 IST 2016
> [INFO] Final Memory: 71M/830M
> [INFO] 
> 
> [ERROR] Failed to execute goal 
> net.alchim31.maven:scala-maven-plugin:3.2.2:compile
> (scala-compile-first) on project spark-sql_2.11: Execution
> scala-compile-first of goal 
> net.alchim31.maven:scala-maven-plugin:3.2.2:compile
> failed. CompileFailed -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the
> -e switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions,
> please read the following articles:
> [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/
> PluginExecutionException
> [ERROR]
> [ERROR] After correcting the problems, you can resume the build with the
> command
> [ERROR]   mvn  -rf :spark-sql_2.11
>
>
>


-- 
Regards,
-- Nachiketa


Re: removing header from csv file

2016-04-27 Thread Nachiketa
Why "without sqlcontext"  ? Could you please describe what is it that you
are trying to accomplish ? Thanks.

Regards,
Nachiketa

On Wed, Apr 27, 2016 at 10:54 AM, Ashutosh Kumar 
wrote:

> I see there is a library spark-csv which can be used for removing header
> and processing of csv files. But it seems it works with sqlcontext only. Is
> there a way to remove header from csv files without sqlcontext ?
>
> Thanks
> Ashutosh
>



-- 
Regards,
-- Nachiketa


Re: Spark 1.4.0, Secure YARN Cluster, Application Master throws 500 connection refused (Resolved)

2015-06-25 Thread Nachiketa
Setting the yarn.resourcemanager.webapp.address.rm1 and
yarn.resourcemanager.webapp.address.rm2 in yarn-site.xml seems to have
resolved the issue.

Appreciate any comments about the regression from 1.3.1 ? Thanks.

Regards,
Nachiketa

On Fri, Jun 26, 2015 at 1:28 AM, Nachiketa 
wrote:

> A few other observations.
>
> 1. Spark 1.3.1 (custom built against HDP 2.2) was running fine against the
> same cluster and same hadoop configuration (hence seems like regression).
>
> 2. HA is enabled for YARN RM and HDFS (not sure if this would impact
> anything but wanted to share anyway).
>
> 3. Found this issue : https://issues.apache.org/jira/browse/SPARK-5837
> and multiple references to other YARN issues to the same. Continuing to
> understand and explore the possibilities documented there.
>
> Regards,
> Nachiketa
>
> On Fri, Jun 26, 2015 at 12:52 AM, Nachiketa 
> wrote:
>
>> Spark 1.4.0 - Custom built from source against Hortonworks HDP 2.2
>> (hadoop 2.6.0+)
>> HDP 2.2 Cluster (Secure, kerberos)
>>
>> spark-shell (--master yarn-client) launches fine and the prompt shows up.
>> Clicking on the Application Master url on the YARN RM UI, throws 500
>> connect error.
>>
>> The same build works well against a non-secure cluster (same HDP
>> distribution).
>>
>> No debug logs or stack trace is easily visible. Where do I look for what
>> is going wrong ? And has anything changed in spark security that could be
>> contributing to this ?
>>
>> Thank you for your help with this.
>>
>>
>> Regards,
>> Nachiketa
>>
>
>
>
> --
> Regards,
> -- Nachiketa
>



-- 
Regards,
-- Nachiketa


Re: Spark 1.4.0, Secure YARN Cluster, Application Master throws 500 connection refused

2015-06-25 Thread Nachiketa
A few other observations.

1. Spark 1.3.1 (custom built against HDP 2.2) was running fine against the
same cluster and same hadoop configuration (hence seems like regression).

2. HA is enabled for YARN RM and HDFS (not sure if this would impact
anything but wanted to share anyway).

3. Found this issue : https://issues.apache.org/jira/browse/SPARK-5837 and
multiple references to other YARN issues to the same. Continuing to
understand and explore the possibilities documented there.

Regards,
Nachiketa

On Fri, Jun 26, 2015 at 12:52 AM, Nachiketa 
wrote:

> Spark 1.4.0 - Custom built from source against Hortonworks HDP 2.2 (hadoop
> 2.6.0+)
> HDP 2.2 Cluster (Secure, kerberos)
>
> spark-shell (--master yarn-client) launches fine and the prompt shows up.
> Clicking on the Application Master url on the YARN RM UI, throws 500
> connect error.
>
> The same build works well against a non-secure cluster (same HDP
> distribution).
>
> No debug logs or stack trace is easily visible. Where do I look for what
> is going wrong ? And has anything changed in spark security that could be
> contributing to this ?
>
> Thank you for your help with this.
>
>
> Regards,
> Nachiketa
>



-- 
Regards,
-- Nachiketa


Spark 1.4.0, Secure YARN Cluster, Application Master throws 500 connection refused

2015-06-25 Thread Nachiketa
Spark 1.4.0 - Custom built from source against Hortonworks HDP 2.2 (hadoop
2.6.0+)
HDP 2.2 Cluster (Secure, kerberos)

spark-shell (--master yarn-client) launches fine and the prompt shows up.
Clicking on the Application Master url on the YARN RM UI, throws 500
connect error.

The same build works well against a non-secure cluster (same HDP
distribution).

No debug logs or stack trace is easily visible. Where do I look for what is
going wrong ? And has anything changed in spark security that could be
contributing to this ?

Thank you for your help with this.


Regards,
Nachiketa