Both from shell with run command and from web client
On Nov 14, 2014 2:32 PM, "Fabian Hueske" <[email protected]> wrote:
>
> In this case, the initialization happens when the InputFormat is
> instantiated at the submission client and the Table info is serialized as
> part of the InputFormat and shipped out to all TaskManagers for execution.
> However, if the initialization is done within configure it happens on each
> TaskManager when initializing the InputFormat.
> These are two separate JVMs in a distributed setting with different
> classpaths.
>
> How do you submit your job for execution?
>
> 2014-11-14 13:58 GMT+01:00 Flavio Pompermaier <[email protected]>:
>
> > The strange thing us that everything works if I create HTable outside
> > configure()..
> > On Nov 14, 2014 10:32 AM, "Stephan Ewen" <[email protected]> wrote:
> >
> > > I think that this is a case where the wrong classloader is used:
> > >
> > > If the HBase classes are part of the flink lib directory, they are
loaded
> > > with the system class loader. When they look for anything in the
> > classpath,
> > > they will do so with the system classloader.
> > >
> > > You configuration is in the user code jar that you submit, so it is
only
> > > available through the user-code classloader.
> > >
> > > Any way you can load the configuration yourself and give that
> > configuration
> > > to HBase?
> > >
> > > Stephan
> > > Am 13.11.2014 22:06 schrieb "Flavio Pompermaier" <[email protected]
>:
> > >
> > > > The only config files available are within the submitted jar. Things
> > > works
> > > > in eclipse using local environment while fails deploying to the
cluster
> > > > On Nov 13, 2014 10:01 PM, <[email protected]> wrote:
> > > >
> > > > > Does the HBase jar in the lib folder contain a config that could
be
> > > used
> > > > > instead of the config in the job jar file? Or is simply no config
at
> > > all
> > > > > available when the configure method is called?
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > --
> > > > > Fabian Hueske
> > > > > Phone: +49 170 5549438
> > > > > Email: [email protected]
> > > > > Web: http://www.user.tu-berlin.de/fabian.hueske
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > From: Flavio Pompermaier
> > > > > Sent: Thursday, 13. November, 2014 21:43
> > > > > To: [email protected]
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > The hbase jar is in the lib directory on each node while the
config
> > > files
> > > > > are within the jar file I submit from the web client.
> > > > > On Nov 13, 2014 9:37 PM, <[email protected]> wrote:
> > > > >
> > > > > > Have you added the hbase.jar file with your HBase config to the
> > ./lib
> > > > > > folders of your Flink setup (JobManager, TaskManager) or is it
> > > bundled
> > > > > with
> > > > > > your job.jar file?
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > --
> > > > > > Fabian Hueske
> > > > > > Phone: +49 170 5549438
> > > > > > Email: [email protected]
> > > > > > Web: http://www.user.tu-berlin.de/fabian.hueske
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > From: Flavio Pompermaier
> > > > > > Sent: Thursday, 13. November, 2014 18:36
> > > > > > To: [email protected]
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > Any help with this? :(
> > > > > >
> > > > > > On Thu, Nov 13, 2014 at 2:06 PM, Flavio Pompermaier <
> > > > > [email protected]>
> > > > > > wrote:
> > > > > >
> > > > > > > We definitely discovered that instantiating HTable and Scan in
> > > > > > configure()
> > > > > > > method of TableInputFormat causes problem in distributed
> > > environment!
> > > > > > > If you look at my implementation at
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
https://github.com/fpompermaier/incubator-flink/blob/master/flink-addons/flink-hbase/src/main/java/org/apache/flink/addons/hbase/TableInputFormat.java
> > > > > > > you can see that Scan and HTable were made transient and
> > recreated
> > > > > within
> > > > > > > configure but this causes HBaseConfiguration.create() to fail
> > > > searching
> > > > > > for
> > > > > > > classpath files...could you help us understanding why?
> > > > > > >
> > > > > > > On Wed, Nov 12, 2014 at 8:10 PM, Flavio Pompermaier <
> > > > > > [email protected]>
> > > > > > > wrote:
> > > > > > >
> > > > > > >> Usually, when I run a mapreduce job both on Spark and Hadoop
I
> > > just
> > > > > put
> > > > > > >> *-site.xml files into the war I submit to the cluster and
that's
> > > > it. I
> > > > > > >> think the problem appeared when I made the HTable a private
> > > > transient
> > > > > > field
> > > > > > >> and the table istantiation was moved in the configure method.
> > > > > > >> Could it be a valid reason? we still have to make a deeper
debug
> > > but
> > > > > I'm
> > > > > > >> trying ro figure out where to investigate..
> > > > > > >> On Nov 12, 2014 8:03 PM, "Robert Metzger" <
[email protected]>
> > > > > wrote:
> > > > > > >>
> > > > > > >>> Hi,
> > > > > > >>> Maybe its an issue with the classpath? As far as I know is
> > Hadoop
> > > > > > reading
> > > > > > >>> the configuration files from the classpath. Maybe is the
> > > > > hbase-site.xml
> > > > > > >>> file not accessible through the classpath when running on
the
> > > > > cluster?
> > > > > > >>>
> > > > > > >>> On Wed, Nov 12, 2014 at 7:40 PM, Flavio Pompermaier <
> > > > > > >>> [email protected]>
> > > > > > >>> wrote:
> > > > > > >>>
> > > > > > >>> > Today we tried tp execute a job on the cluster instead of
on
> > > > local
> > > > > > >>> executor
> > > > > > >>> > and we faced the problem that the hbase-site.xml was
> > basically
> > > > > > >>> ignored. Is
> > > > > > >>> > there a reason why the TableInputFormat is working
correctly
> > on
> > > > > local
> > > > > > >>> > environment while it doesn't on a cluster?
> > > > > > >>> > On Nov 10, 2014 10:56 AM, "Fabian Hueske" <
> > [email protected]>
> > > > > > wrote:
> > > > > > >>> >
> > > > > > >>> > > I don't think we need to bundle the HBase input and
output
> > > > format
> > > > > > in
> > > > > > >>> a
> > > > > > >>> > > single PR.
> > > > > > >>> > > So, I think we can proceed with the IF only and target
the
> > OF
> > > > > > later.
> > > > > > >>> > > However, the fix for Kryo should be in the master before
> > > > merging
> > > > > > the
> > > > > > >>> PR.
> > > > > > >>> > > Till is currently working on that and said he expects
this
> > to
> > > > be
> > > > > > >>> done by
> > > > > > >>> > > end of the week.
> > > > > > >>> > >
> > > > > > >>> > > Cheers, Fabian
> > > > > > >>> > >
> > > > > > >>> > >
> > > > > > >>> > > 2014-11-07 12:49 GMT+01:00 Flavio Pompermaier <
> > > > > > [email protected]
> > > > > > >>> >:
> > > > > > >>> > >
> > > > > > >>> > > > I fixed also the profile for Cloudera CDH5.1.3. You
can
> > > build
> > > > > it
> > > > > > >>> with
> > > > > > >>> > the
> > > > > > >>> > > > command:
> > > > > > >>> > > > mvn clean install -Dmaven.test.skip=true
> > > > > -Dhadoop.profile=2
> > > > > > >>> > > > -Pvendor-repos,cdh5.1.3
> > > > > > >>> > > >
> > > > > > >>> > > > However, it would be good to generate the specific jar
> > when
> > > > > > >>> > > > releasing..(e.g.
> > > > > > >>> > > >
> > flink-addons:flink-hbase:0.8.0-hadoop2-cdh5.1.3-incubating)
> > > > > > >>> > > >
> > > > > > >>> > > > Best,
> > > > > > >>> > > > Flavio
> > > > > > >>> > > >
> > > > > > >>> > > > On Fri, Nov 7, 2014 at 12:44 PM, Flavio Pompermaier <
> > > > > > >>> > > [email protected]>
> > > > > > >>> > > > wrote:
> > > > > > >>> > > >
> > > > > > >>> > > > > I've just updated the code on my fork (synch with
> > current
> > > > > > master
> > > > > > >>> and
> > > > > > >>> > > > > applied improvements coming from comments on related
> > PR).
> > > > > > >>> > > > > I still have to understand how to write results
back to
> > > an
> > > > > > HBase
> > > > > > >>> > > > > Sink/OutputFormat...
> > > > > > >>> > > > >
> > > > > > >>> > > > >
> > > > > > >>> > > > > On Mon, Nov 3, 2014 at 12:05 PM, Flavio Pompermaier
<
> > > > > > >>> > > > [email protected]>
> > > > > > >>> > > > > wrote:
> > > > > > >>> > > > >
> > > > > > >>> > > > >> Thanks for the detailed answer. So if I run a job
from
> > > my
> > > > > > >>> machine
> > > > > > >>> > I'll
> > > > > > >>> > > > >> have to download all the scanned data in a
> > table..right?
> > > > > > >>> > > > >>
> > > > > > >>> > > > >> Always regarding the GenericTableOutputFormat it is
> > not
> > > > > clear
> > > > > > >>> to me
> > > > > > >>> > > how
> > > > > > >>> > > > >> to proceed..
> > > > > > >>> > > > >> I saw in the hadoop compatibility addon that it is
> > > > possible
> > > > > to
> > > > > > >>> have
> > > > > > >>> > > such
> > > > > > >>> > > > >> compatibility using HBaseUtils class so the open
> > method
> > > > > should
> > > > > > >>> > become
> > > > > > >>> > > > >> something like:
> > > > > > >>> > > > >>
> > > > > > >>> > > > >> @Override
> > > > > > >>> > > > >> public void open(int taskNumber, int numTasks)
throws
> > > > > > >>> IOException {
> > > > > > >>> > > > >> if (Integer.toString(taskNumber + 1).length() > 6)
{
> > > > > > >>> > > > >> throw new IOException("Task id too large.");
> > > > > > >>> > > > >> }
> > > > > > >>> > > > >> TaskAttemptID taskAttemptID =
> > > > > > >>> > TaskAttemptID.forName("attempt__0000_r_"
> > > > > > >>> > > > >> + String.format("%" + (6 -
> > Integer.toString(taskNumber +
> > > > > > >>> > 1).length())
> > > > > > >>> > > +
> > > > > > >>> > > > >> "s"," ").replace(" ", "0")
> > > > > > >>> > > > >> + Integer.toString(taskNumber + 1)
> > > > > > >>> > > > >> + "_0");
> > > > > > >>> > > > >> this.configuration.set("mapred.task.id",
> > > > > > >>> > taskAttemptID.toString());
> > > > > > >>> > > > >> this.configuration.setInt("mapred.task.partition",
> > > > > taskNumber
> > > > > > +
> > > > > > >>> 1);
> > > > > > >>> > > > >> // for hadoop 2.2
> > > > > > >>> > > > >> this.configuration.set("mapreduce.task.attempt.id",
> > > > > > >>> > > > >> taskAttemptID.toString());
> > > > > > >>> > > > >>
this.configuration.setInt("mapreduce.task.partition",
> > > > > > >>> taskNumber +
> > > > > > >>> > 1);
> > > > > > >>> > > > >> try {
> > > > > > >>> > > > >> this.context =
> > > > > > >>> > > > >>
> > > > > HadoopUtils.instantiateTaskAttemptContext(this.configuration,
> > > > > > >>> > > > >> taskAttemptID);
> > > > > > >>> > > > >> } catch (Exception e) {
> > > > > > >>> > > > >> throw new RuntimeException(e);
> > > > > > >>> > > > >> }
> > > > > > >>> > > > >> final HFileOutputFormat2 outFormat = new
> > > > > HFileOutputFormat2();
> > > > > > >>> > > > >> try {
> > > > > > >>> > > > >> this.writer =
outFormat.getRecordWriter(this.context);
> > > > > > >>> > > > >> } catch (InterruptedException iex) {
> > > > > > >>> > > > >> throw new IOException("Opening the writer was
> > > > interrupted.",
> > > > > > >>> iex);
> > > > > > >>> > > > >> }
> > > > > > >>> > > > >> }
> > > > > > >>> > > > >>
> > > > > > >>> > > > >> But I'm not sure about how to pass the JobConf to
the
> > > > class,
> > > > > > if
> > > > > > >>> to
> > > > > > >>> > > merge
> > > > > > >>> > > > >> config fileas, where HFileOutputFormat2 writes the
> > data
> > > > and
> > > > > > how
> > > > > > >>> to
> > > > > > >>> > > > >> implement the public void writeRecord(Record
record)
> > > API.
> > > > > > >>> > > > >> Could I do a little chat off the mailing list with
the
> > > > > > >>> implementor
> > > > > > >>> > of
> > > > > > >>> > > > >> this extension?
> > > > > > >>> > > > >>
> > > > > > >>> > > > >> On Mon, Nov 3, 2014 at 11:51 AM, Fabian Hueske <
> > > > > > >>> [email protected]>
> > > > > > >>> > > > >> wrote:
> > > > > > >>> > > > >>
> > > > > > >>> > > > >>> Hi Flavio
> > > > > > >>> > > > >>>
> > > > > > >>> > > > >>> let me try to answer your last question on the
user's
> > > > list
> > > > > > (to
> > > > > > >>> the
> > > > > > >>> > > best
> > > > > > >>> > > > >>> of
> > > > > > >>> > > > >>> my HBase knowledge).
> > > > > > >>> > > > >>> "I just wanted to known if and how regiom
splitting
> > is
> > > > > > >>> handled. Can
> > > > > > >>> > > you
> > > > > > >>> > > > >>> explain me in detail how Flink and HBase
works?what
> > is
> > > > not
> > > > > > >>> fully
> > > > > > >>> > > clear
> > > > > > >>> > > > to
> > > > > > >>> > > > >>> me is when computation is done by region servers
and
> > > when
> > > > > > data
> > > > > > >>> > start
> > > > > > >>> > > > flow
> > > > > > >>> > > > >>> to a Flink worker (that in ky test job is only my
pc)
> > > and
> > > > > how
> > > > > > >>> ro
> > > > > > >>> > > > >>> undertsand
> > > > > > >>> > > > >>> better the important logged info to understand if
my
> > > job
> > > > is
> > > > > > >>> > > performing
> > > > > > >>> > > > >>> well"
> > > > > > >>> > > > >>>
> > > > > > >>> > > > >>> HBase partitions its tables into so called
"regions"
> > of
> > > > > keys
> > > > > > >>> and
> > > > > > >>> > > stores
> > > > > > >>> > > > >>> the
> > > > > > >>> > > > >>> regions distributed in the cluster using HDFS. I
> > think
> > > an
> > > > > > HBase
> > > > > > >>> > > region
> > > > > > >>> > > > >>> can
> > > > > > >>> > > > >>> be thought of as a HDFS block. To make reading an
> > HBase
> > > > > table
> > > > > > >>> > > > efficient,
> > > > > > >>> > > > >>> region reads should be locally done, i.e., an
> > > InputFormat
> > > > > > >>> should
> > > > > > >>> > > > >>> primarily
> > > > > > >>> > > > >>> read region that are stored on the same machine as
> > the
> > > IF
> > > > > is
> > > > > > >>> > running
> > > > > > >>> > > > on.
> > > > > > >>> > > > >>> Flink's InputSplits partition the HBase input by
> > > regions
> > > > > and
> > > > > > >>> add
> > > > > > >>> > > > >>> information about the storage location of the
region.
> > > > > During
> > > > > > >>> > > execution,
> > > > > > >>> > > > >>> input splits are assigned to InputFormats that
can do
> > > > local
> > > > > > >>> reads.
> > > > > > >>> > > > >>>
> > > > > > >>> > > > >>> Best, Fabian
> > > > > > >>> > > > >>>
> > > > > > >>> > > > >>> 2014-11-03 11:13 GMT+01:00 Stephan Ewen <
> > > > [email protected]
> > > > > >:
> > > > > > >>> > > > >>>
> > > > > > >>> > > > >>> > Hi!
> > > > > > >>> > > > >>> >
> > > > > > >>> > > > >>> > The way of passing parameters through the
> > > configuration
> > > > > is
> > > > > > >>> very
> > > > > > >>> > old
> > > > > > >>> > > > >>> (the
> > > > > > >>> > > > >>> > original HBase format dated back to that time).
I
> > > would
> > > > > > >>> simply
> > > > > > >>> > make
> > > > > > >>> > > > the
> > > > > > >>> > > > >>> > HBase format take those parameters through the
> > > > > constructor.
> > > > > > >>> > > > >>> >
> > > > > > >>> > > > >>> > Greetings,
> > > > > > >>> > > > >>> > Stephan
> > > > > > >>> > > > >>> >
> > > > > > >>> > > > >>> >
> > > > > > >>> > > > >>> > On Mon, Nov 3, 2014 at 10:59 AM, Flavio
> > Pompermaier <
> > > > > > >>> > > > >>> [email protected]>
> > > > > > >>> > > > >>> > wrote:
> > > > > > >>> > > > >>> >
> > > > > > >>> > > > >>> > > The problem is that I also removed the
> > > > > > >>> GenericTableOutputFormat
> > > > > > >>> > > > >>> because
> > > > > > >>> > > > >>> > > there is an incompatibility between hadoop1
and
> > > > hadoop2
> > > > > > for
> > > > > > >>> > class
> > > > > > >>> > > > >>> > > TaskAttemptContext and
TaskAttemptContextImpl..
> > > > > > >>> > > > >>> > > then it would be nice if the user doesn't
have to
> > > > worry
> > > > > > >>> about
> > > > > > >>> > > > passing
> > > > > > >>> > > > >>> > > pact.hbase.jtkey and pact.job.id parameters..
> > > > > > >>> > > > >>> > > I think it is probably a good idea to remove
> > > hadoop1
> > > > > > >>> > > compatibility
> > > > > > >>> > > > >>> and
> > > > > > >>> > > > >>> > keep
> > > > > > >>> > > > >>> > > enable HBase addon only for hadoop2 (as
before)
> > and
> > > > > > decide
> > > > > > >>> how
> > > > > > >>> > to
> > > > > > >>> > > > >>> mange
> > > > > > >>> > > > >>> > > those 2 parameters..
> > > > > > >>> > > > >>> > >
> > > > > > >>> > > > >>> > > On Mon, Nov 3, 2014 at 10:19 AM, Stephan Ewen
<
> > > > > > >>> > [email protected]>
> > > > > > >>> > > > >>> wrote:
> > > > > > >>> > > > >>> > >
> > > > > > >>> > > > >>> > > > It is fine to remove it, in my opinion.
> > > > > > >>> > > > >>> > > >
> > > > > > >>> > > > >>> > > > On Mon, Nov 3, 2014 at 10:11 AM, Flavio
> > > > Pompermaier <
> > > > > > >>> > > > >>> > > [email protected]>
> > > > > > >>> > > > >>> > > > wrote:
> > > > > > >>> > > > >>> > > >
> > > > > > >>> > > > >>> > > > > That is one class I removed because it was
> > > using
> > > > > the
> > > > > > >>> > > deprecated
> > > > > > >>> > > > >>> API
> > > > > > >>> > > > >>> > > > > GenericDataSink..I can restore them but
the
> > it
> > > > will
> > > > > > be
> > > > > > >>> a
> > > > > > >>> > good
> > > > > > >>> > > > >>> idea to
> > > > > > >>> > > > >>> > > > > remove those warning (also because from
what
> > I
> > > > > > >>> understood
> > > > > > >>> > the
> > > > > > >>> > > > >>> Record
> > > > > > >>> > > > >>> > > APIs
> > > > > > >>> > > > >>> > > > > are going to be removed).
> > > > > > >>> > > > >>> > > > >
> > > > > > >>> > > > >>> > > > > On Mon, Nov 3, 2014 at 9:51 AM, Fabian
> > Hueske <
> > > > > > >>> > > > >>> [email protected]>
> > > > > > >>> > > > >>> > > > wrote:
> > > > > > >>> > > > >>> > > > >
> > > > > > >>> > > > >>> > > > > > I'm not familiar with the HBase
connector
> > > code,
> > > > > but
> > > > > > >>> are
> > > > > > >>> > you
> > > > > > >>> > > > >>> maybe
> > > > > > >>> > > > >>> > > > looking
> > > > > > >>> > > > >>> > > > > > for the GenericTableOutputFormat?
> > > > > > >>> > > > >>> > > > > >
> > > > > > >>> > > > >>> > > > > > 2014-11-03 9:44 GMT+01:00 Flavio
> > Pompermaier
> > > <
> > > > > > >>> > > > >>> [email protected]
> > > > > > >>> > > > >>> > >:
> > > > > > >>> > > > >>> > > > > >
> > > > > > >>> > > > >>> > > > > > > | was trying to modify the example
> > setting
> > > > > > >>> > > > hbaseDs.output(new
> > > > > > >>> > > > >>> > > > > > > HBaseOutputFormat()); but I can't see
any
> > > > > > >>> > > HBaseOutputFormat
> > > > > > >>> > > > >>> > > > > class..maybe
> > > > > > >>> > > > >>> > > > > > we
> > > > > > >>> > > > >>> > > > > > > shall use another class?
> > > > > > >>> > > > >>> > > > > > >
> > > > > > >>> > > > >>> > > > > > > On Mon, Nov 3, 2014 at 9:39 AM, Flavio
> > > > > > Pompermaier
> > > > > > >>> <
> > > > > > >>> > > > >>> > > > > [email protected]
> > > > > > >>> > > > >>> > > > > > >
> > > > > > >>> > > > >>> > > > > > > wrote:
> > > > > > >>> > > > >>> > > > > > >
> > > > > > >>> > > > >>> > > > > > > > Maybe that's something I could add
to
> > the
> > > > > HBase
> > > > > > >>> > example
> > > > > > >>> > > > and
> > > > > > >>> > > > >>> > that
> > > > > > >>> > > > >>> > > > > could
> > > > > > >>> > > > >>> > > > > > be
> > > > > > >>> > > > >>> > > > > > > > better documented in the Wiki.
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > > > Since we're talking about the
wiki..I
> > was
> > > > > > >>> looking at
> > > > > > >>> > > the
> > > > > > >>> > > > >>> Java
> > > > > > >>> > > > >>> > > API (
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > >
> > > > > > >>> > > > >>> > > > > >
> > > > > > >>> > > > >>> > > > >
> > > > > > >>> > > > >>> > > >
> > > > > > >>> > > > >>> > >
> > > > > > >>> > > > >>> >
> > > > > > >>> > > > >>>
> > > > > > >>> > > >
> > > > > > >>> > >
> > > > > > >>> >
> > > > > > >>>
> > > > > >
> > > > >
> > > >
> > >
> >
http://flink.incubator.apache.org/docs/0.6-incubating/java_api_guide.html
> > > > > > >>> > > > >>> )
> > > > > > >>> > > > >>> > > > > > > > and the link to the KMeans example
is
> > not
> > > > > > working
> > > > > > >>> > > (where
> > > > > > >>> > > > it
> > > > > > >>> > > > >>> > says
> > > > > > >>> > > > >>> > > > For
> > > > > > >>> > > > >>> > > > > a
> > > > > > >>> > > > >>> > > > > > > > complete example program, have a
look
> > at
> > > > > KMeans
> > > > > > >>> > > > Algorithm).
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > > > Best,
> > > > > > >>> > > > >>> > > > > > > > Flavio
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > > > On Mon, Nov 3, 2014 at 9:12 AM,
Flavio
> > > > > > >>> Pompermaier <
> > > > > > >>> > > > >>> > > > > > [email protected]
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > > > wrote:
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > > >> Ah ok, perfect! That was the reason
> > why
> > > I
> > > > > > >>> removed it
> > > > > > >>> > > :)
> > > > > > >>> > > > >>> > > > > > > >>
> > > > > > >>> > > > >>> > > > > > > >> On Mon, Nov 3, 2014 at 9:10 AM,
> > Stephan
> > > > > Ewen <
> > > > > > >>> > > > >>> > [email protected]>
> > > > > > >>> > > > >>> > > > > > wrote:
> > > > > > >>> > > > >>> > > > > > > >>
> > > > > > >>> > > > >>> > > > > > > >>> You do not really need a HBase
data
> > > sink.
> > > > > You
> > > > > > >>> can
> > > > > > >>> > > call
> > > > > > >>> > > > >>> > > > > > > >>> "DataSet.output(new
> > > > > > >>> > > > >>> > > > > > > >>> HBaseOutputFormat())"
> > > > > > >>> > > > >>> > > > > > > >>>
> > > > > > >>> > > > >>> > > > > > > >>> Stephan
> > > > > > >>> > > > >>> > > > > > > >>> Am 02.11.2014 23:05 schrieb
"Flavio
> > > > > > >>> Pompermaier" <
> > > > > > >>> > > > >>> > > > > > [email protected]
> > > > > > >>> > > > >>> > > > > > > >:
> > > > > > >>> > > > >>> > > > > > > >>>
> > > > > > >>> > > > >>> > > > > > > >>> > Just one last thing..I removed
the
> > > > > > >>> HbaseDataSink
> > > > > > >>> > > > >>> because I
> > > > > > >>> > > > >>> > > > think
> > > > > > >>> > > > >>> > > > > it
> > > > > > >>> > > > >>> > > > > > > was
> > > > > > >>> > > > >>> > > > > > > >>> > using the old APIs..can someone
> > help
> > > me
> > > > > in
> > > > > > >>> > updating
> > > > > > >>> > > > >>> that
> > > > > > >>> > > > >>> > > class?
> > > > > > >>> > > > >>> > > > > > > >>> >
> > > > > > >>> > > > >>> > > > > > > >>> > On Sun, Nov 2, 2014 at 10:55 AM,
> > > Flavio
> > > > > > >>> > > Pompermaier <
> > > > > > >>> > > > >>> > > > > > > >>> [email protected]>
> > > > > > >>> > > > >>> > > > > > > >>> > wrote:
> > > > > > >>> > > > >>> > > > > > > >>> >
> > > > > > >>> > > > >>> > > > > > > >>> > > Indeed this time the build has
> > been
> > > > > > >>> successful
> > > > > > >>> > :)
> > > > > > >>> > > > >>> > > > > > > >>> > >
> > > > > > >>> > > > >>> > > > > > > >>> > > On Sun, Nov 2, 2014 at 10:29
AM,
> > > > Fabian
> > > > > > >>> Hueske
> > > > > > >>> > <
> > > > > > >>> > > > >>> > > > > > [email protected]
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > > >>> > wrote:
> > > > > > >>> > > > >>> > > > > > > >>> > >
> > > > > > >>> > > > >>> > > > > > > >>> > >> You can also setup Travis to
> > build
> > > > > your
> > > > > > >>> own
> > > > > > >>> > > Github
> > > > > > >>> > > > >>> > > > > repositories
> > > > > > >>> > > > >>> > > > > > by
> > > > > > >>> > > > >>> > > > > > > >>> > linking
> > > > > > >>> > > > >>> > > > > > > >>> > >> it to your Github account.
That
> > > way
> > > > > > >>> Travis can
> > > > > > >>> > > > >>> build all
> > > > > > >>> > > > >>> > > > your
> > > > > > >>> > > > >>> > > > > > > >>> branches
> > > > > > >>> > > > >>> > > > > > > >>> > >> (and
> > > > > > >>> > > > >>> > > > > > > >>> > >> you can also trigger
rebuilds if
> > > > > > something
> > > > > > >>> > > fails).
> > > > > > >>> > > > >>> > > > > > > >>> > >> Not sure if we can manually
> > > trigger
> > > > > > >>> retrigger
> > > > > > >>> > > > >>> builds on
> > > > > > >>> > > > >>> > > the
> > > > > > >>> > > > >>> > > > > > Apache
> > > > > > >>> > > > >>> > > > > > > >>> > >> repository.
> > > > > > >>> > > > >>> > > > > > > >>> > >>
> > > > > > >>> > > > >>> > > > > > > >>> > >> Support for Hadoop 1 and 2 is
> > > > indeed a
> > > > > > >>> very
> > > > > > >>> > good
> > > > > > >>> > > > >>> > addition
> > > > > > >>> > > > >>> > > > :-)
> > > > > > >>> > > > >>> > > > > > > >>> > >>
> > > > > > >>> > > > >>> > > > > > > >>> > >> For the discusion about the
PR
> > > > > itself, I
> > > > > > >>> would
> > > > > > >>> > > > need
> > > > > > >>> > > > >>> a
> > > > > > >>> > > > >>> > bit
> > > > > > >>> > > > >>> > > > more
> > > > > > >>> > > > >>> > > > > > > time
> > > > > > >>> > > > >>> > > > > > > >>> to
> > > > > > >>> > > > >>> > > > > > > >>> > >> become more familiar with
> > HBase. I
> > > > do
> > > > > > >>> also not
> > > > > > >>> > > > have
> > > > > > >>> > > > >>> a
> > > > > > >>> > > > >>> > > HBase
> > > > > > >>> > > > >>> > > > > > setup
> > > > > > >>> > > > >>> > > > > > > >>> > >> available
> > > > > > >>> > > > >>> > > > > > > >>> > >> here.
> > > > > > >>> > > > >>> > > > > > > >>> > >> Maybe somebody else of the
> > > community
> > > > > who
> > > > > > >>> was
> > > > > > >>> > > > >>> involved
> > > > > > >>> > > > >>> > > with a
> > > > > > >>> > > > >>> > > > > > > >>> previous
> > > > > > >>> > > > >>> > > > > > > >>> > >> version of the HBase
connector
> > > could
> > > > > > >>> comment
> > > > > > >>> > on
> > > > > > >>> > > > your
> > > > > > >>> > > > >>> > > > question.
> > > > > > >>> > > > >>> > > > > > > >>> > >>
> > > > > > >>> > > > >>> > > > > > > >>> > >> Best, Fabian
> > > > > > >>> > > > >>> > > > > > > >>> > >>
> > > > > > >>> > > > >>> > > > > > > >>> > >> 2014-11-02 9:57 GMT+01:00
Flavio
> > > > > > >>> Pompermaier <
> > > > > > >>> > > > >>> > > > > > > [email protected]
> > > > > > >>> > > > >>> > > > > > > >>> >:
> > > > > > >>> > > > >>> > > > > > > >>> > >>
> > > > > > >>> > > > >>> > > > > > > >>> > >> > As suggestes by Fabian I
moved
> > > the
> > > > > > >>> > discussion
> > > > > > >>> > > on
> > > > > > >>> > > > >>> this
> > > > > > >>> > > > >>> > > > > mailing
> > > > > > >>> > > > >>> > > > > > > >>> list.
> > > > > > >>> > > > >>> > > > > > > >>> > >> >
> > > > > > >>> > > > >>> > > > > > > >>> > >> > I think that what is still
to
> > be
> > > > > > >>> discussed
> > > > > > >>> > is
> > > > > > >>> > > > >>> how to
> > > > > > >>> > > > >>> > > > > > retrigger
> > > > > > >>> > > > >>> > > > > > > >>> the
> > > > > > >>> > > > >>> > > > > > > >>> > >> build
> > > > > > >>> > > > >>> > > > > > > >>> > >> > on Travis (I don't have an
> > > > account)
> > > > > > and
> > > > > > >>> if
> > > > > > >>> > the
> > > > > > >>> > > > PR
> > > > > > >>> > > > >>> can
> > > > > > >>> > > > >>> > be
> > > > > > >>> > > > >>> > > > > > > >>> integrated.
> > > > > > >>> > > > >>> > > > > > > >>> > >> >
> > > > > > >>> > > > >>> > > > > > > >>> > >> > Maybe what I can do is to
move
> > > the
> > > > > > HBase
> > > > > > >>> > > example
> > > > > > >>> > > > >>> in
> > > > > > >>> > > > >>> > the
> > > > > > >>> > > > >>> > > > test
> > > > > > >>> > > > >>> > > > > > > >>> package
> > > > > > >>> > > > >>> > > > > > > >>> > >> (right
> > > > > > >>> > > > >>> > > > > > > >>> > >> > now I left it in the main
> > > folder)
> > > > so
> > > > > > it
> > > > > > >>> will
> > > > > > >>> > > > force
> > > > > > >>> > > > >>> > > Travis
> > > > > > >>> > > > >>> > > > to
> > > > > > >>> > > > >>> > > > > > > >>> rebuild.
> > > > > > >>> > > > >>> > > > > > > >>> > >> > I'll do it within a couple
of
> > > > hours.
> > > > > > >>> > > > >>> > > > > > > >>> > >> >
> > > > > > >>> > > > >>> > > > > > > >>> > >> > Another thing I forgot to
say
> > is
> > > > > that
> > > > > > >>> the
> > > > > > >>> > > hbase
> > > > > > >>> > > > >>> > > extension
> > > > > > >>> > > > >>> > > > is
> > > > > > >>> > > > >>> > > > > > now
> > > > > > >>> > > > >>> > > > > > > >>> > >> compatible
> > > > > > >>> > > > >>> > > > > > > >>> > >> > with both hadoop 1 and 2.
> > > > > > >>> > > > >>> > > > > > > >>> > >> >
> > > > > > >>> > > > >>> > > > > > > >>> > >> > Best,
> > > > > > >>> > > > >>> > > > > > > >>> > >> > Flavio
> > > > > > >>> > > > >>> > > > > > > >>> > >>
> > > > > > >>> > > > >>> > > > > > > >>> > >
> > > > > > >>> > > > >>> > > > > > > >>> >
> > > > > > >>> > > > >>> > > > > > > >>>
> > > > > > >>> > > > >>> > > > > > > >>
> > > > > > >>> > > > >>> > > > > > > >
> > > > > > >>> > > > >>> > > > > > >
> > > > > > >>> > > > >>> > > > > >
> > > > > > >>> > > > >>> > > > >
> > > > > > >>> > > > >>> > > >
> > > > > > >>> > > > >>> > >
> > > > > > >>> > > > >>> >
> > > > > > >>> > > > >>>
> > > > > > >>> > > > >>
> > > > > > >>> > > > >>
> > > > > > >>> > > > >>
> > > > > > >>> > > > >
> > > > > > >>> > > >
> > > > > > >>> > >
> > > > > > >>> >
> > > > > > >>>
> > > > > > >>
> > > > > > >
> > > > > > >
> > > >
> > >
> >