On Tue, Jan 22, 2013 at 1:48 PM, Amila Suriarachchi <[email protected]> wrote:

>
>
> On Tue, Jan 22, 2013 at 12:56 PM, Lalaji Sureshika <[email protected]>wrote:
>
>> Hi,
>>
>> On Tue, Jan 22, 2013 at 11:19 AM, Amila Suriarachchi <[email protected]>wrote:
>>
>>>
>>>
>>> On Tue, Jan 22, 2013 at 12:14 PM, Buddhika Chamith 
>>> <[email protected]>wrote:
>>>
>>>> This usually happens when an event not conforming to the stream
>>>> definition registered is sent. May be there has been some change to stream
>>>> definition in the toolbox or at agent side in versions being used?
>>>>
>>>
>>> I ran with some default settings. I'll check with a MySQL DB.
>>>
>>
>> One additional note.When I invoke the API with authentication level as
>> none,I could see stats on some graphs as in attached image,but for two
>> graphs,I'm not getting data and below error[1] shown in BAM side.
>>
>> When going through the code again,at the Authentication Handler
>> level,when the authentication level set to 'NONE' for an API, the username
>> of authentication context has set as null.Because of that at the Usage
>> Handler level,we are getting the username from Authentication Context to
>> pass with the event as null and thus there seems the sending event is not
>> conforming to the stream definition as Buddhika mentioned.
>>
>
> We should be able to define those fields as null in Trift IDL and handle
> Cassendra, Hive levels, reports accordingly.
>

Another option is to use the user name as anonymous if the user is not
available.

thanks,
Amila.

>
> thanks,
> Amila.
>
>
>>
>> SanjeewaM is currently looking in to this..
>>
>> Thanks;
>>
>> [1]  TID: [0] [BAM] [2013-01-22 10:06:13,488] ERROR
>> {org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation} -  Failed to
>> write data to database
>> {org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation}
>> org.h2.jdbc.JdbcSQLException: NULL not allowed for column "USERID"; SQL
>> statement:
>> INSERT INTO API_REQUEST_SUMMARY
>> (context,version,consumerkey,total_request_count,userid,max_request_time,api_version,api)
>> VALUES (?,?,?,?,?,?,?,?) [90006-140]
>>  at org.h2.message.DbException.getJdbcSQLException(DbException.java:327)
>> at org.h2.message.DbException.get(DbException.java:167)
>>  at org.h2.message.DbException.get(DbException.java:144)
>> at org.h2.table.Column.validateConvertUpdateSequence(Column.java:294)
>>  at org.h2.table.Table.validateConvertUpdateSequence(Table.java:621)
>> at org.h2.command.dml.Insert.insertRows(Insert.java:116)
>>  at org.h2.command.dml.Insert.update(Insert.java:82)
>> at org.h2.command.CommandContainer.update(CommandContainer.java:70)
>>  at org.h2.command.Command.executeUpdate(Command.java:199)
>> at
>> org.h2.jdbc.JdbcPreparedStatement.executeUpdateInternal(JdbcPreparedStatement.java:141)
>>  at
>> org.h2.jdbc.JdbcPreparedStatement.executeUpdate(JdbcPreparedStatement.java:127)
>> at
>> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.insertData(DBOperation.java:141)
>>  at
>> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBOperation.writeToDB(DBOperation.java:62)
>> at
>> org.wso2.carbon.hadoop.hive.jdbc.storage.db.DBRecordWriter.write(DBRecordWriter.java:35)
>>  at
>> org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:589)
>> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
>>  at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762)
>> at
>> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
>>  at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
>> at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762)
>>  at
>> org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
>> at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:471)
>>  at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:762)
>> at
>> org.apache.hadoop.hive.ql.exec.GroupByOperator.forward(GroupByOperator.java:959)
>>  at
>> org.apache.hadoop.hive.ql.exec.GroupByOperator.closeOp(GroupByOperator.java:1012)
>> at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:557)
>>  at
>> org.apache.hadoop.hive.ql.exec.ExecReducer.close(ExecReducer.java:303)
>> at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:528)
>>  at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:419)
>> at
>> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:256)
>> TID: [0] [BAM] [2013-01-22 10:06:13,497] ERROR {ExecReducer} -  Hit error
>> while closing operators - failing tree {ExecReducer}
>> TID: [0] [BAM] [2013-01-22 10:06:14,466] ERROR
>> {org.apache.hadoop.hive.ql.exec.ExecDriver} -  Ended Job = job_local_0001
>> with errors
>>
>>
>>>
>>> thanks,
>>> Amila.
>>>
>>>
>>>>
>>>> Regards
>>>> Buddhika
>>>>
>>>>
>>>> On Tue, Jan 22, 2013 at 12:07 PM, Lalaji Sureshika <[email protected]>wrote:
>>>>
>>>>>  extract data from the incoming request to the gateway,without
>>>>> depending on the security scheme attached to the particular API resource
>>>>> verb..
>>>>> I tried same scenario with keeping the security level as none for a
>>>>> particular API resource and without subscribe it to any app.I could able 
>>>>> to
>>>>> view stats from publisher side.[AM 1.3.0 and BAM 2.0.1]
>>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> *Amila Suriarachchi*
>>>
>>> Software Architect
>>> WSO2 Inc. ; http://wso2.com
>>> lean . enterprise . middleware
>>>
>>> phone : +94 71 3082805
>>>
>>
>>
>>
>> --
>> Lalaji Sureshika
>> Software Engineer; Development Technologies Team;WSO2, Inc.;
>> http://wso2.com/
>> email: [email protected]; cell: +94 71 608 6811
>> blog: http://lalajisureshika.blogspot.com
>>
>>
>>
>
>
> --
> *Amila Suriarachchi*
>
> Software Architect
> WSO2 Inc. ; http://wso2.com
> lean . enterprise . middleware
>
> phone : +94 71 3082805
>



-- 
*Amila Suriarachchi*

Software Architect
WSO2 Inc. ; http://wso2.com
lean . enterprise . middleware

phone : +94 71 3082805
_______________________________________________
Dev mailing list
[email protected]
http://wso2.org/cgi-bin/mailman/listinfo/dev

Reply via email to