Will this be the same implementation as session.read.jdbc("") and then call
this code continuously like how we are running HUDI in continuous mode?

On Mon, Sep 16, 2019 at 9:09 PM Vinoth Chandar <[email protected]> wrote:

> Thanks, Taher! Any takers for driving this? This is something I would be
> very interested in getting involved with. Dont have the bandwidth atm :/
>
> On Sun, Sep 15, 2019 at 11:15 PM Taher Koitawala <[email protected]>
> wrote:
>
> > Thank you all for your support. JIRA filed at
> > https://issues.apache.org/jira/browse/HUDI-251
> >
> > Regards,
> > Taher Koitawala
> >
> > On Mon, Sep 16, 2019 at 11:34 AM Taher Koitawala <[email protected]>
> > wrote:
> >
> > > Since everyone is fully onboard. I am creating a JIRA to track this.
> > >
> > > On Sun, Sep 15, 2019 at 9:47 AM [email protected] <[email protected]
> >
> > > wrote:
> > >
> > >>
> > >> +1. Agree with everyone's point. Go for it Taher !!
> > >> Balaji.V    On Saturday, September 14, 2019, 07:44:04 PM PDT, Bhavani
> > >> Sudha Saktheeswaran <[email protected]> wrote:
> > >>
> > >>  +1 I  think adding new sources to DeltaStreamer is really valuable.
> > >>
> > >> Thanks,
> > >> Sudha
> > >>
> > >> On Sat, Sep 14, 2019 at 7:52 AM vino yang <[email protected]>
> > wrote:
> > >>
> > >> > Hi Taher,
> > >> >
> > >> > IMO, it's a good supplement to Hudi.
> > >> >
> > >> > So +1 from my side.
> > >> >
> > >> > Vinoth Chandar <[email protected]> 于2019年9月14日周六 下午10:23写道:
> > >> >
> > >> > > Hi Taher,
> > >> > >
> > >> > > I am fully onboard on this. This is such a frequently asked
> question
> > >> and
> > >> > > having it all doable with a simple DeltaStreamer command would be
> > >> really
> > >> > > powerful.
> > >> > >
> > >> > > +1
> > >> > >
> > >> > > - Vinoth
> > >> > >
> > >> > > On 2019/09/14 05:51:05, Taher Koitawala <[email protected]>
> wrote:
> > >> > > > Hi All,
> > >> > > >          Currently, we are trying to pull data incrementally
> from
> > >> our
> > >> > > RDBMS
> > >> > > > sources, however the way we are doing this is with HUDI is to
> > >> create a
> > >> > > > spark table on top of the JDBC source using [1] which writes raw
> > >> data
> > >> > to
> > >> > > an
> > >> > > > HDFS dir. We then use DeltaStreamer dfs-source to write that to
> a
> > >> HUDI
> > >> > > > upsert COPY_ON_WRITE table.
> > >> > > >
> > >> > > >          However, I think it would be really helpful in such use
> > >> cases
> > >> > > > where DeltaStreamer had something like a JDBC-source instead of
> > >> sqoop
> > >> > or
> > >> > > > temp tables and then we could leave that in a continuous mode
> > with a
> > >> > > > timestamp column and an interval which allows us to express how
> > >> > > frequently
> > >> > > > DeltaStreamer should check for new updates or inserts on RDBMS.
> > >> > > >
> > >> > > > 1: CREATE TABLE mysql_temp_table
> > >> > > > USING org.apache.spark.sql.jdbc
> > >> > > > OPTIONS (
> > >> > > >      url  "jdbc:mysql://
> > >> > > >
> > >> > >
> > >> >
> > >>
> >
> https://urldefense.proofpoint.com/v2/url?u=http-3A__data.source.mysql.com&d=DwIFaQ&c=r2dcLCtU9q6n0vrtnDw9vg&r=oyPDRKU5b-LuEWWyf8gacx4mFFydIGdyS50OKdxizX0&m=kd2JZkFO9u_nWk8s__l1rNlfZ0cQ_zXOjURNBNmmJo4&s=zIAG-Ct3xm-8XBHg7Gv4mxPF7YpQJ5wxWTarYnJlJDE&e=
> > >> >
> > >>
> >
> :3306/database?user=mysql_user&password=password&zeroDateTimeBehavior=CONVERT_TO_NULL
> > >> > > > ",
> > >> > > >      dbtable "database.table_name",
> > >> > > >      fetchSize "1000000",
> > >> > > >      partitionColumn "contact_id", lowerBound "1",
> > >> > > > upperBound "2962429",
> > >> > > > numPartitions "62"
> > >> > > > );
> > >> > > >
> > >> > > > Regards,
> > >> > > > Taher Koitawala
> > >> > > >
> > >> > >
> > >> >
> > >
> > >
> >
>

Reply via email to