Hi Gokul,
I tried it with fresh pack and used default hbase config. Now it's working
fine. Thanks for the help. Since I'm running hbase standalone mode, I used
following datasource configuration:
<datasource>
<name>WSO2_ANALYTICS_RS_DB_HBASE</name>
<description>The datasource used for analytics file
system</description>
<jndiConfig>
<name>jdbc/WSO2HBaseDB</name>
</jndiConfig>
<definition type="HBASE">
<configuration>
<property>
<name>hbase.master</name>
<value>localhost:16000</value>
</property>
<!-- <property>
<name>fs.hdfs.impl</name>
<value>org.apache.hadoop.hdfs.DistributedFileSystem</value>
</property>
<property>
<name>fs.file.impl</name>
<value>org.apache.hadoop.fs.LocalFileSystem</value>
</property> -->
</configuration>
</definition>
</datasource>
Thanks.
On Wed, Sep 9, 2015 at 3:07 PM, Thanuja Uruththirakodeeswaran <
[email protected]> wrote:
> Hi Gokul,
>
> I've added this configuration and tried again. But getting the same error.
>
> Thanks.
>
> On Wed, Sep 9, 2015 at 2:50 PM, Gokul Balakrishnan <[email protected]> wrote:
>
>> Hi Thanuja,
>>
>> As given in the thread "[DAS 3.0.0 Beta] No FileSystem for scheme: file",
>> can you add the following properties and try:
>>
>> <property>
>> <name>fs.hdfs.impl</name>
>> <value>org.apache.hadoop.hdfs.DistributedFileSystem</value>
>> </property>
>> <property>
>> <name>fs.file.impl</name>
>> <value>org.apache.hadoop.fs.LocalFileSystem</value>
>> </property>
>>
>> These settings are available out-of-the-box in the sample datasource
>> configuration in recent DAS packs, any reason why they went missing?
>>
>> Thanks,
>>
>> On 9 September 2015 at 14:47, Thanuja Uruththirakodeeswaran <
>> [email protected]> wrote:
>>
>>> Hi Gokul,
>>>
>>> I'm running Hbase in *standalone mode* and used HBase as eventsotre in
>>> DAS. II tried with the following two datasource definitions:
>>>
>>> 1.
>>> <datasource>
>>> <name>WSO2_ANALYTICS_EVENT_STORE_DB_HBASE</name>
>>> <description>The datasource used for analytics file
>>> system</description>
>>> <jndiConfig>
>>> <name>jdbc/WSO2HBaseDB</name>
>>> </jndiConfig>
>>> <definition type="HBASE">
>>> <configuration>
>>> <property>
>>> <name>hbase.master</name>
>>> <value>localhost:16000</value>
>>> </property>
>>> </configuration>
>>> </definition>
>>> </datasource>
>>>
>>> 2.
>>> <datasource>
>>> <name>WSO2_ANALYTICS_EVENT_STORE_DB_HBASE</name>
>>> <description>The datasource used for analytics file
>>> system</description>
>>> <jndiConfig>
>>> <name>jdbc/WSO2HBaseDB</name>
>>> </jndiConfig>
>>> <definition type="HBASE">
>>> <configuration>
>>> <property>
>>> <name>hbase.master</name>
>>> <value>localhost:16000</value>
>>> </property>
>>> <property>
>>> <name>hbase.zookeeper.quorum</name>
>>> <value>127.0.0.1</value>
>>> </property>
>>> <property>
>>> <name>hbase.zookeeper.property.clientPort</name>
>>> <value>2181</value>
>>> </property>
>>> </configuration>
>>> </definition>
>>> </datasource>
>>>
>>> I'm getting the following error:
>>>
>>> TID: [-1234] [] [2015-09-09 14:35:53,954] WARN
>>> {org.apache.hadoop.hbase.util.DynamicClassLoader} - Failed to identify the
>>> fs of dir
>>> /home/thanujau/Desktop/wso2das-3.0.0-SNAPSHOT/tmp/hbase-thanujau/hbase/lib,
>>> ignored {org.apache.hadoop.hbase.util.DynamicClassLoader}
>>> java.io.IOException: No FileSystem for scheme: file
>>> at
>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2584)
>>> at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2591)
>>> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:91)
>>> at
>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2630)
>>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2612)
>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:370)
>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:169)
>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:354)
>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296)
>>> at
>>> org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:104)
>>> at
>>> org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:229)
>>> at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64)
>>> at
>>> org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75)
>>> at
>>> org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:86)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.retrieveClusterId(ConnectionManager.java:833)
>>> at
>>> org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.<init>(ConnectionManager.java:623)
>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>>> at
>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>>> at
>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>>
>>>
>>> *I've attached the full log. I used the same configuration with beta
>>> pack. It worked fine. But in the latest pack I'm getting this error.*
>>>
>>> *Also I logged into hbase shell and there is no probelm.*
>>>
>>> Thanks.
>>>
>>> --
>>> Thanuja Uruththirakodeeswaran
>>> Software Engineer
>>> WSO2 Inc.;http://wso2.com
>>> lean.enterprise.middleware
>>>
>>> mobile: +94 774363167
>>>
>>
>>
>>
>> --
>> Gokul Balakrishnan
>> Senior Software Engineer,
>> WSO2, Inc. http://wso2.com
>> Mob: +94 77 593 5789 | +1 650 272 9927
>>
>
>
>
> --
> Thanuja Uruththirakodeeswaran
> Software Engineer
> WSO2 Inc.;http://wso2.com
> lean.enterprise.middleware
>
> mobile: +94 774363167
>
--
Thanuja Uruththirakodeeswaran
Software Engineer
WSO2 Inc.;http://wso2.com
lean.enterprise.middleware
mobile: +94 774363167
_______________________________________________
Dev mailing list
[email protected]
http://wso2.org/cgi-bin/mailman/listinfo/dev