Joe, the first hurdle is to get ojdbc6.jar downloaded and installed in
/usr/share/java.  There's a link created in /usr/hdp/2.3.0.0-2557/hive/lib/
but points to nothing.

Here's the hurdle I can't get past.  If you install and run kite-dataset
from the web site and run through the example with debug and verbose turned
on (below) you get the output below.  It thinks mapreduce.tar.gz doesn't
exist, but it does (way down below).  I've run this as users root and hdfs
with no joy.  Thanks for looking.

debug=true ./kite-dataset -v csv-import sandwiches.csv sandwiches

WARNING: Use "yarn jar" to launch YARN applications.
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/usr/hdp/2.3.0.0-2557/hadoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/usr/hdp/2.3.0.0-2557/zookeeper/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
1 job failure(s) occurred:
org.kitesdk.tools.CopyTask:
Kite(dataset:file:/tmp/0c1454eb-7831-4d6b-85a2-63a6cc8c51... ID=1 (1/1)(1):
java.io.FileNotFoundException: File
file:/hdp/apps/2.3.0.0-2557/mapreduce/mapreduce.tar.gz does not exist
    at
org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:606)
    at
org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:819)
    at
org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:596)
    at
org.apache.hadoop.fs.DelegateToFileSystem.getFileStatus(DelegateToFileSystem.java:110)
    at
org.apache.hadoop.fs.AbstractFileSystem.resolvePath(AbstractFileSystem.java:467)
    at org.apache.hadoop.fs.FilterFs.resolvePath(FilterFs.java:157)
    at org.apache.hadoop.fs.FileContext$25.next(FileContext.java:2193)
    at org.apache.hadoop.fs.FileContext$25.next(FileContext.java:2189)
    at org.apache.hadoop.fs.FSLinkResolver.resolve(FSLinkResolver.java:90)
    at org.apache.hadoop.fs.FileContext.resolve(FileContext.java:2189)
    at org.apache.hadoop.fs.FileContext.resolvePath(FileContext.java:601)
    at
org.apache.hadoop.mapreduce.JobSubmitter.addMRFrameworkToDistributedCache(JobSubmitter.java:457)
    at
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:142)
    at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
    at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:415)
    at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
    at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
    at
org.apache.crunch.hadoop.mapreduce.lib.jobcontrol.CrunchControlledJob.submit(CrunchControlledJob.java:329)
    at
org.apache.crunch.hadoop.mapreduce.lib.jobcontrol.CrunchJobControl.startReadyJobs(CrunchJobControl.java:204)
    at
org.apache.crunch.hadoop.mapreduce.lib.jobcontrol.CrunchJobControl.pollJobStatusAndStartNewOnes(CrunchJobControl.java:238)
    at
org.apache.crunch.impl.mr.exec.MRExecutor.monitorLoop(MRExecutor.java:112)
    at
org.apache.crunch.impl.mr.exec.MRExecutor.access$000(MRExecutor.java:55)
    at org.apache.crunch.impl.mr.exec.MRExecutor$1.run(MRExecutor.java:83)
    at java.lang.Thread.run(Thread.java:745)

[hdfs@sandbox ~]$ hdfs dfs -ls /hdp/apps/2.3.0.0-2557/mapreduce
Found 2 items
-r--r--r--   1 hdfs hadoop     105893 2015-08-20 08:36
/hdp/apps/2.3.0.0-2557/mapreduce/hadoop-streaming.jar
-r--r--r--   1 hdfs hadoop  207888607 2015-08-20 08:33
/hdp/apps/2.3.0.0-2557/mapreduce/mapreduce.tar.gz


On Thu, Oct 15, 2015 at 3:22 PM, Joe Witt <[email protected]> wrote:

> Chris,
>
> Are you seeing errors in NiFi or in HDP?  If you're seeing errors in
> NiFi can you please send us the logs?
>
> Thanks
> Joe
>
> On Thu, Oct 15, 2015 at 3:02 PM, Christopher Wilson <[email protected]>
> wrote:
> > Has anyone gotten Kite to work on HDP?  I'd wanted to do this very thing
> but
> > am running into all kinds of issues with having .jar files not in the
> > distributed cache (basically in /apps/hdp).
> >
> > Any feedback appreciated.
> >
> > -Chris
> >
> > On Sat, Sep 19, 2015 at 11:04 AM, Tyler Hawkes <[email protected]>
> > wrote:
> >>
> >> Thanks for the link. I'm using
> >> "dataset:hive://hadoop01:9083/default/sandwiches". hadoop01 has hive on
> it.
> >>
> >> On Fri, Sep 18, 2015 at 7:36 AM Jeff <[email protected]> wrote:
> >>>
> >>> Not sure if this is what you are looking for but it has a bit on kite.
> >>>
> >>> http://ingest.tips/2014/12/22/getting-started-with-apache-nifi/
> >>>
> >>> -cb
> >>>
> >>>
> >>> On Sep 18, 2015, at 8:32 AM, Bryan Bende <[email protected]> wrote:
> >>>
> >>> Hi Tyler,
> >>>
> >>> Unfortunately I don't think there are any tutorials on this. Can you
> >>> provide an example of the dataset uri you specified that is showing as
> >>> invalid?
> >>>
> >>> Thanks,
> >>>
> >>> Bryan
> >>>
> >>> On Fri, Sep 18, 2015 at 12:36 AM, Tyler Hawkes <[email protected]
> >
> >>> wrote:
> >>>>
> >>>> I'm just getting going on NiFi and trying to write data to Hive either
> >>>> from Kafka or an RDBMS. After setting up the hadoop configuration
> files and
> >>>> a target dataset uri it says the uri is invalid. I'm wondering if
> there's a
> >>>> tutorial on getting kite set up with my version of hive (HDP 2.2
> running
> >>>> hive 0.14) and nifi since I've been unable to find anything on google
> or on
> >>>> the mailing list archive and the documentation of StoreInKiteDataset
> it
> >>>> lacking a lot of detail.
> >>>>
> >>>> Any help on this would be greatly appreciated.
> >>>
> >>>
> >>>
> >
>

Reply via email to