Was the pack smoke tested? Are the BAM scenarios working?

Sent from my Samsung Galaxy S3
On May 11, 2013 7:59 PM, "Sanjeewa Malalgoda" <[email protected]> wrote:

> Hi,
> I think this is because error in data source configuration in BAM side.
> Please double check WSO2BAM_CASSANDRA_DATASOURCE configuration. It should
> be as follows(have to set port offset there as well). It worked for me.
>
>         <datasource>
>             <name>WSO2BAM_CASSANDRA_DATASOURCE</name>
>             <description>The datasource used for Cassandra
> data</description>
>             <definition type="RDBMS">
>                 <configuration>
>                     <url>jdbc:cassandra://127.0.0.1:9161/EVENT_KS</url>
>                     <username>admin</username>
>                     <password>admin</password>
>                 </configuration>
>             </definition>
>         </datasource>
>
> Thanks,
> Sanjeewa.
>
> On Sat, May 11, 2013 at 3:01 PM, Nuwan Dias <[email protected]> wrote:
>
>> Hi,
>>
>> I am trying out the latest BAM pack to monitor statistics of the API
>> Gateway. But I'm getting the following error on the BAM console when the
>> Hive script runs. What could be wrong here? The issue is a Thrift
>> connection refused error. I am running an API manager (offset 0), ESB
>> (offset 2) and BAM (offset 1) on the same machine. I can see that the data
>> has been written to Cassandra. I have attached the toolbox as well.
>>
>> Hive history
>> file=/home/nuwan/Work/AM/10-05-2013/wso2bam-2.3.0/tmp/hive/wso2-querylogs/hive_job_log_nuwan_201305111446_1870617742.txt
>> [2013-05-11 14:48:00,092] ERROR
>> {org.apache.hadoop.hive.cassandra.CassandraProxyClient} -  Error while
>> trying to connect to cassandra host:localhost
>> org.apache.hadoop.hive.cassandra.CassandraException: unable to connect to
>> server
>> at
>> org.apache.hadoop.hive.cassandra.CassandraClientHolder.initClient(CassandraClientHolder.java:57)
>>  at
>> org.apache.hadoop.hive.cassandra.CassandraClientHolder.<init>(CassandraClientHolder.java:37)
>> at
>> org.apache.hadoop.hive.cassandra.CassandraProxyClient.createConnection(CassandraProxyClient.java:181)
>>  at
>> org.apache.hadoop.hive.cassandra.CassandraProxyClient.initializeConnection(CassandraProxyClient.java:207)
>> at
>> org.apache.hadoop.hive.cassandra.CassandraProxyClient.<init>(CassandraProxyClient.java:144)
>>  at
>> org.apache.hadoop.hive.cassandra.CassandraManager.openConnection(CassandraManager.java:185)
>> at
>> org.apache.hadoop.hive.cassandra.CassandraStorageHandler.preCreateTable(CassandraStorageHandler.java:224)
>>  at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:397)
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:540)
>>  at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3479)
>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:225)
>>  at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:133)
>> at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)
>>  at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1351)
>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1126)
>>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:934)
>> at
>> org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:201)
>>  at
>> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:187)
>> at
>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:337)
>>  at
>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:232)
>> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>>  at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>> at
>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>>  at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>> at java.lang.Thread.run(Thread.java:662)
>> Caused by: org.apache.thrift.transport.TTransportException:
>> java.net.ConnectException: Connection refused
>> at org.apache.thrift.transport.TSocket.open(TSocket.java:183)
>>  at
>> org.apache.thrift.transport.TFramedTransport.open(TFramedTransport.java:81)
>> at
>> org.apache.hadoop.hive.cassandra.CassandraClientHolder.initClient(CassandraClientHolder.java:54)
>>  ... 24 more
>> Caused by: java.net.ConnectException: Connection refused
>> at java.net.PlainSocketImpl.socketConnect(Native Method)
>>  at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:351)
>> at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:213)
>>  at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:200)
>> at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
>>  at java.net.Socket.connect(Socket.java:529)
>> at org.apache.thrift.transport.TSocket.open(TSocket.java:178)
>>  ... 26 more
>> FAILED: Error in metadata: java.lang.NullPointerException
>> [2013-05-11 14:48:00,093] ERROR {org.apache.hadoop.hive.ql.exec.Task} -
>>  FAILED: Error in metadata: java.lang.NullPointerException
>> org.apache.hadoop.hive.ql.metadata.HiveException:
>> java.lang.NullPointerException
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:546)
>>  at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3479)
>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:225)
>>  at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:133)
>> at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57)
>>  at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1351)
>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1126)
>>  at org.apache.hadoop.hive.ql.Driver.run(Driver.java:934)
>> at
>> org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:201)
>>  at
>> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:187)
>> at
>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:337)
>>  at
>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:232)
>> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>>  at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>> at
>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>>  at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>> at java.lang.Thread.run(Thread.java:662)
>> Caused by: java.lang.NullPointerException
>> at
>> org.apache.hadoop.hive.cassandra.CassandraProxyClient.initializeConnection(CassandraProxyClient.java:223)
>>  at
>> org.apache.hadoop.hive.cassandra.CassandraProxyClient.<init>(CassandraProxyClient.java:144)
>> at
>> org.apache.hadoop.hive.cassandra.CassandraManager.openConnection(CassandraManager.java:185)
>>  at
>> org.apache.hadoop.hive.cassandra.CassandraStorageHandler.preCreateTable(CassandraStorageHandler.java:224)
>> at
>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:397)
>>  at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:540)
>> ... 16 more
>>
>> FAILED: Execution Error, return code 1 from
>> org.apache.hadoop.hive.ql.exec.DDLTask
>> [2013-05-11 14:48:00,094] ERROR {org.apache.hadoop.hive.ql.Driver} -
>>  FAILED: Execution Error, return code 1 from
>> org.apache.hadoop.hive.ql.exec.DDLTask
>> [2013-05-11 14:48:00,094] ERROR
>> {org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl} -  Error
>> while executing Hive script.
>> Query returned non-zero code: 9, cause: FAILED: Execution Error, return
>> code 1 from org.apache.hadoop.hive.ql.exec.DDLTask
>> java.sql.SQLException: Query returned non-zero code: 9, cause: FAILED:
>> Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask
>>  at
>> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:189)
>> at
>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:337)
>>  at
>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:232)
>> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>>  at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>> at
>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>>  at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>> at java.lang.Thread.run(Thread.java:662)
>> [2013-05-11 14:48:00,096] ERROR
>> {org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask} -  Error while
>> executing script : am_stats_analyzer_664
>> org.wso2.carbon.analytics.hive.exception.HiveExecutionException: Error
>> while executing Hive script.Query returned non-zero code: 9, cause: FAILED:
>> Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask
>>  at
>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl.execute(HiveExecutorServiceImpl.java:117)
>> at
>> org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask.execute(HiveScriptExecutorTask.java:60)
>>  at
>> org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute(TaskQuartzJobAdapter.java:56)
>> at org.quartz.core.JobRunShell.run(JobRunShell.java:213)
>>  at
>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
>> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>>  at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>> at
>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>>  at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>> at java.lang.Thread.run(Thread.java:662)
>>
>> Thanks,
>> NuwanD
>>
>>
>> On Fri, May 10, 2013 at 11:09 AM, Gihan Anuruddha <[email protected]> wrote:
>>
>>> [Adding dev@]
>>> Hi Nuwan,
>>>
>>> Noted. Sorry for the inconvenience.
>>>
>>> Regards,
>>> Gihan
>>>
>>>
>>> On Fri, May 10, 2013 at 10:53 AM, Nuwan Dias <[email protected]> wrote:
>>>
>>>> Hi Gihan,
>>>>
>>>> This should be in the dev@ list.
>>>>
>>>> Thanks,
>>>> NuwanD.
>>>>
>>>>
>>>> On Fri, May 10, 2013 at 10:31 AM, Gihan Anuruddha <[email protected]>wrote:
>>>>
>>>>> Hi QA Team,
>>>>>
>>>>> Please find latest BAM 2.3.0 pack for QA testing at [1].
>>>>>
>>>>> [1] -
>>>>> http://173.164.178.33/builds/BAM-2.3.0/09-05-2013/wso2bam-2.3.0.zip
>>>>>
>>>>> Regards,
>>>>> Gihan
>>>>>
>>>>> --
>>>>> ---
>>>>> W.G. Gihan Anuruddha
>>>>> Senior Software Engineer | WSO2, Inc.
>>>>> M: +94772272595
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Nuwan Dias
>>>>
>>>> Member, Management Committee - Solutions Technology Group
>>>> Software Engineer - WSO2, Inc. http://wso2.com
>>>> email : [email protected]
>>>> Phone : +94 777 775 729
>>>>
>>>
>>>
>>>
>>> --
>>> ---
>>> W.G. Gihan Anuruddha
>>> Senior Software Engineer | WSO2, Inc.
>>> M: +94772272595
>>>
>>
>>
>>
>> --
>> Nuwan Dias
>>
>> Member, Management Committee - Solutions Technology Group
>> Software Engineer - WSO2, Inc. http://wso2.com
>> email : [email protected]
>> Phone : +94 777 775 729
>>
>> _______________________________________________
>> Dev mailing list
>> [email protected]
>> http://wso2.org/cgi-bin/mailman/listinfo/dev
>>
>>
>
>
> --
> *Sanjeewa Malalgoda*
> WSO2 Inc.
> Mobile : +14084122175 | +94713068779
>
>  <http://sanjeewamalalgoda.blogspot.com/>blog
> :http://sanjeewamalalgoda.blogspot.com/<http://sanjeewamalalgoda.blogspot.com/>
>
> _______________________________________________
> Dev mailing list
> [email protected]
> http://wso2.org/cgi-bin/mailman/listinfo/dev
>
>
_______________________________________________
Dev mailing list
[email protected]
http://wso2.org/cgi-bin/mailman/listinfo/dev

Reply via email to