[
https://issues.apache.org/jira/browse/HBASE-18175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16053466#comment-16053466
]
Yi Liang edited comment on HBASE-18175 at 6/19/17 4:22 AM:
-----------------------------------------------------------
Hi Sean,
In the new patch, I add a submodule called hbase-spark-it under hbase-spark
module, the parent of hbase-spark-it is hbase-spark, not hbase, this can make
hbase-spark module more independent.
Now the hbase-spark looks like
- hbase-spark
-hbase-spark-it
-src
-pom
-src
-pom
And I also have some thoughts on HBase-16179, only solve that jira then we can
push hbase-spark into hbase-2.0
And I saw your comments in that jira, there might be some requirement like, we
need to consider spark-1.6, spark-2.0, and also scala 2.10, scala 2.11
I remember I try both the scala 2.10 and 2.11 for hbase-spark, it seems that
there is no code change for scala version change; the code differences are only
in different spark versions.
I think we can have hbase-spark module structure as below
- hbase-spark
-hbase-spark-1.6
-src (diff code for 1.6)
-pom
-hbase-spark-2.0
-hbase-spark-it
-src
-pom
And user can change scala version on their own for both spark1.6 and spark2.0,
What do you think of this idea?
Thanks
was (Author: easyliangjob):
Hi Sean,
In the new patch, I add a submodule called hbase-spark-it under hbase-spark
module, the parent of hbase-spark-it is hbase-spark, not hbase, this can make
hbase-spark module more independent.
Now the hbase-spark looks like
- hbase-spark
-hbase-spark-it
-src
-pom
-src
-pom
And I also have some thoughts on HBase-16179, only solve that jira then we can
push hbase-spark into hbase-2.0
And I saw your comments in that jira, there might be some requirement like, we
need to consider spark-1.6, spark-2.0, and also scala 2.10, scala 2.11
I remember I try both the scala 2.10 and 2.11 for hbase-spark, it seems that
there is no code change for scala version change; the code differences are only
in different spark versions.
I think we can have hbase-spark module structure as below
- hbase-spark
-hbase-spark-1.6
-src (diff code for 1.6)
-pom
-hbase-spark-2.0
-hbase-spark-it
-src
-pom
And user can change scala version on their own for both spark1.6 and spark2.0,
What do you think of this idea?
Thanks
> Add hbase-spark integration test into hbase-it
> ----------------------------------------------
>
> Key: HBASE-18175
> URL: https://issues.apache.org/jira/browse/HBASE-18175
> Project: HBase
> Issue Type: Test
> Components: spark
> Reporter: Yi Liang
> Assignee: Yi Liang
> Priority: Critical
> Fix For: 2.0.0
>
> Attachments: hbase-18175-v1.patch
>
>
> After HBASE-17574, all test under hbase-spark are regarded as unit test, and
> this jira will add integration test of hbase-spark into hbase-it. This patch
> run same tests as mapreduce.IntegrationTestBulkLoad, just change mapreduce to
> spark.
> test in Maven:
> mvn verify -Dit.test=IntegrationTestSparkBulkLoad
> test on cluster:
> spark-submit --class
> org.apache.hadoop.hbase.spark.IntegrationTestSparkBulkLoad
> HBASE_HOME/lib/hbase-it-2.0.0-SNAPSHOT-tests.jar
> -Dhbase.spark.bulkload.chainlength=500000 -m slowDeterministic
--
This message was sent by Atlassian JIRA
(v6.4.14#64029)