Hi Ted,

You can check full stack trace log from the attachment at Jira:

https://issues.apache.org/jira/browse/GORA-386

Kind Regards,
Furkan KAMACI

On Wed, Aug 26, 2015 at 6:55 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> My understanding is that people on this mailing list who are interested to
> help can log comments on the GORA JIRA.
> HBase integration with Spark is proven to work. So the intricacies should
> be on Gora side.
>
> On Wed, Aug 26, 2015 at 8:08 AM, Furkan KAMACI <furkankam...@gmail.com>
> wrote:
>
>> Btw, here is the source code of GoraInputFormat.java :
>>
>>
>> https://github.com/kamaci/gora/blob/master/gora-core/src/main/java/org/apache/gora/mapreduce/GoraInputFormat.java
>> 26 Ağu 2015 18:05 tarihinde "Furkan KAMACI" <furkankam...@gmail.com>
>> yazdı:
>>
>> I'll send an e-mail to Gora dev list too and also attach my patch into my
>>> GSoC Jira issue you mentioned and then we can continue at there.
>>>
>>> Before I do that stuff, I wanted to get Spark dev community's ideas to
>>> solve my problem due to you may have faced such kind of problems before.
>>> 26 Ağu 2015 17:13 tarihinde "Ted Yu" <yuzhih...@gmail.com> yazdı:
>>>
>>>> I found GORA-386 Gora Spark Backend Support
>>>>
>>>> Should the discussion be continued there ?
>>>>
>>>> Cheers
>>>>
>>>> On Wed, Aug 26, 2015 at 7:02 AM, Ted Malaska <ted.mala...@cloudera.com>
>>>> wrote:
>>>>
>>>>> Where is the input format class.  When every I use the search on your
>>>>> github it says "We couldn’t find any issues matching 'GoraInputFormat'"
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Aug 26, 2015 at 9:48 AM, Furkan KAMACI <furkankam...@gmail.com
>>>>> > wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> Here is the MapReduceTestUtils.testSparkWordCount()
>>>>>>
>>>>>>
>>>>>> https://github.com/kamaci/gora/blob/master/gora-core/src/test/java/org/apache/gora/mapreduce/MapReduceTestUtils.java#L108
>>>>>>
>>>>>> Here is SparkWordCount
>>>>>>
>>>>>>
>>>>>> https://github.com/kamaci/gora/blob/8f1acc6d4ef6c192e8fc06287558b7bc7c39b040/gora-core/src/examples/java/org/apache/gora/examples/spark/SparkWordCount.java
>>>>>>
>>>>>> Lastly, here is GoraSparkEngine:
>>>>>>
>>>>>>
>>>>>> https://github.com/kamaci/gora/blob/master/gora-core/src/main/java/org/apache/gora/spark/GoraSparkEngine.java
>>>>>>
>>>>>> Kind Regards,
>>>>>> Furkan KAMACI
>>>>>>
>>>>>> On Wed, Aug 26, 2015 at 4:40 PM, Ted Malaska <
>>>>>> ted.mala...@cloudera.com> wrote:
>>>>>>
>>>>>>> Where can I find the code for MapReduceTestUtils.testSparkWordCount?
>>>>>>>
>>>>>>> On Wed, Aug 26, 2015 at 9:29 AM, Furkan KAMACI <
>>>>>>> furkankam...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> Here is the test method I've ignored due to Connection Refused
>>>>>>>> problem failure:
>>>>>>>>
>>>>>>>>
>>>>>>>> https://github.com/kamaci/gora/blob/master/gora-hbase/src/test/java/org/apache/gora/hbase/mapreduce/TestHBaseStoreWordCount.java#L65
>>>>>>>>
>>>>>>>> I've implemented a Spark backend for Apache Gora as GSoC project
>>>>>>>> and this is the latest obstacle that I should solve. If you can help 
>>>>>>>> me,
>>>>>>>> you are welcome.
>>>>>>>>
>>>>>>>> Kind Regards,
>>>>>>>> Furkan KAMACI
>>>>>>>>
>>>>>>>> On Wed, Aug 26, 2015 at 3:45 PM, Ted Malaska <
>>>>>>>> ted.mala...@cloudera.com> wrote:
>>>>>>>>
>>>>>>>>> I've always used HBaseTestingUtility and never really had much
>>>>>>>>> trouble. I use that for all my unit testing between Spark and HBase.
>>>>>>>>>
>>>>>>>>> Here are some code examples if your interested
>>>>>>>>>
>>>>>>>>> --Main HBase-Spark Module
>>>>>>>>> https://github.com/apache/hbase/tree/master/hbase-spark
>>>>>>>>>
>>>>>>>>> --Unit test that cover all basic connections
>>>>>>>>>
>>>>>>>>> https://github.com/apache/hbase/blob/master/hbase-spark/src/test/scala/org/apache/hadoop/hbase/spark/HBaseContextSuite.scala
>>>>>>>>>
>>>>>>>>> --If you want to look at the old stuff before it went into HBase
>>>>>>>>> https://github.com/cloudera-labs/SparkOnHBase
>>>>>>>>>
>>>>>>>>> Let me know if that helps
>>>>>>>>>
>>>>>>>>> On Wed, Aug 26, 2015 at 5:40 AM, Ted Yu <yuzhih...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Can you log the contents of the Configuration you pass from Spark
>>>>>>>>>> ?
>>>>>>>>>> The output would give you some clue.
>>>>>>>>>>
>>>>>>>>>> Cheers
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Aug 26, 2015, at 2:30 AM, Furkan KAMACI <
>>>>>>>>>> furkankam...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>> Hi Ted,
>>>>>>>>>>
>>>>>>>>>> I'll check Zookeeper connection but another test method which
>>>>>>>>>> runs on hbase without Spark works without any error. Hbase version is
>>>>>>>>>> 0.98.8-hadoop2 and I use Spark 1.3.1
>>>>>>>>>>
>>>>>>>>>> Kind Regards,
>>>>>>>>>> Furkan KAMACI
>>>>>>>>>> 26 Ağu 2015 12:08 tarihinde "Ted Yu" <yuzhih...@gmail.com> yazdı:
>>>>>>>>>>
>>>>>>>>>>> The connection failure was to zookeeper.
>>>>>>>>>>>
>>>>>>>>>>> Have you verified that localhost:2181 can serve requests ?
>>>>>>>>>>> What version of hbase was Gora built against ?
>>>>>>>>>>>
>>>>>>>>>>> Cheers
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Aug 26, 2015, at 1:50 AM, Furkan KAMACI <
>>>>>>>>>>> furkankam...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>> Hi,
>>>>>>>>>>>
>>>>>>>>>>> I start an Hbase cluster for my test class. I use that helper
>>>>>>>>>>> class:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> https://github.com/apache/gora/blob/master/gora-hbase/src/test/java/org/apache/gora/hbase/util/HBaseClusterSingleton.java
>>>>>>>>>>>
>>>>>>>>>>> and use it as like that:
>>>>>>>>>>>
>>>>>>>>>>> private static final HBaseClusterSingleton cluster =
>>>>>>>>>>> HBaseClusterSingleton.build(1);
>>>>>>>>>>>
>>>>>>>>>>> I retrieve configuration object as follows:
>>>>>>>>>>>
>>>>>>>>>>> cluster.getConf()
>>>>>>>>>>>
>>>>>>>>>>> and I use it at Spark as follows:
>>>>>>>>>>>
>>>>>>>>>>> sparkContext.newAPIHadoopRDD(conf, MyInputFormat.class, clazzK,
>>>>>>>>>>>     clazzV);
>>>>>>>>>>>
>>>>>>>>>>> When I run my test there is no need to startup an Hbase cluster
>>>>>>>>>>> because Spark will connect to my dummy cluster. However when I run 
>>>>>>>>>>> my test
>>>>>>>>>>> method it throws an error:
>>>>>>>>>>>
>>>>>>>>>>> 2015-08-26 01:19:59,558 INFO [Executor task launch
>>>>>>>>>>> worker-0-SendThread(localhost:2181)] zookeeper.ClientCnxn
>>>>>>>>>>> (ClientCnxn.java:logStartConnect(966)) - Opening socket connection 
>>>>>>>>>>> to
>>>>>>>>>>> server localhost/127.0.0.1:2181. Will not attempt to
>>>>>>>>>>> authenticate using SASL (unknown error)
>>>>>>>>>>>
>>>>>>>>>>> 2015-08-26 01:19:59,559 WARN [Executor task launch
>>>>>>>>>>> worker-0-SendThread(localhost:2181)] zookeeper.ClientCnxn
>>>>>>>>>>> (ClientCnxn.java:run(1089)) - Session 0x0 for server null, 
>>>>>>>>>>> unexpected
>>>>>>>>>>> error, closing socket connection and attempting reconnect
>>>>>>>>>>> java.net.ConnectException: Connection refused at
>>>>>>>>>>> sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) at
>>>>>>>>>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739)
>>>>>>>>>>>  at
>>>>>>>>>>> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350)
>>>>>>>>>>> at 
>>>>>>>>>>> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068)
>>>>>>>>>>> Hbase tests, which do not run on Spark, works well. When I check
>>>>>>>>>>> the logs I see that cluster and Spark is started up correctly:
>>>>>>>>>>>
>>>>>>>>>>> 2015-08-26 01:35:21,791 INFO [main] hdfs.MiniDFSCluster
>>>>>>>>>>> (MiniDFSCluster.java:waitActive(2055)) - Cluster is active
>>>>>>>>>>>
>>>>>>>>>>> 2015-08-26 01:35:40,334 INFO [main] util.Utils
>>>>>>>>>>> (Logging.scala:logInfo(59)) - Successfully started service 
>>>>>>>>>>> 'sparkDriver' on
>>>>>>>>>>> port 56941.
>>>>>>>>>>> I realized that when I start up an hbase from command line my
>>>>>>>>>>> test method for Spark connects to it!
>>>>>>>>>>>
>>>>>>>>>>> So, does it means that it doesn't care about the conf I passed
>>>>>>>>>>> to it? Any ideas about how to solve it?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>

Reply via email to