Hi,
Thaks for the reply. I added the property. But flume cannot write a log
file from linux to HDFS.


On Thu, Aug 16, 2012 at 2:25 PM, Mohammad Tariq <[email protected]> wrote:

> Hello Sandeep
>
>    Please edit the "dfs.web.ugi" property in hdfs-site.xml. It is by
> default "webuser,webgroup".
>
> Regards,
>     Mohammad Tariq
>
>
>
> On Thu, Aug 16, 2012 at 11:50 PM, Sandeep Reddy P <
> [email protected]> wrote:
>
>> Hi,
>> I'm using flume-ng to write data from log file to HDFS but unable to
>> write. I got the following exceptions in Name node logs
>>
>> 2012-08-16 13:57:42,560 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>> transactions: 1 Total time for transactions(ms): 0Number of transactions
>> batched in Syncs: 0 Number of syncs: 0 SyncTimes(ms): 0 0 0
>> 2012-08-16 13:57:59,293 WARN
>> org.apache.hadoop.security.ShellBasedUnixGroupsMapping: got exception
>> trying to get groups for user webuser
>> org.apache.hadoop.util.Shell$ExitCodeException: id: webuser: No such user
>>
>>         at org.apache.hadoop.util.Shell.runCommand(Shell.java:255)
>>         at org.apache.hadoop.util.Shell.run(Shell.java:182)
>>         at
>> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:375)
>>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:461)
>>         at org.apache.hadoop.util.Shell.execCommand(Shell.java:444)
>>         at
>> org.apache.hadoop.security.ShellBasedUnixGroupsMapping.getUnixGroups(ShellBasedUnixGroupsMapping.java:66)
>>         at
>> org.apache.hadoop.security.ShellBasedUnixGroupsMapping.getGroups(ShellBasedUnixGroupsMapping.java:43)
>>         at org.apache.hadoop.security.Groups.getGroups(Groups.java:79)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.getGroupNames(UserGroupInformation.java:1083)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.<init>(FSPermissionChecker.java:50)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:5203)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkTraverse(FSNamesystem.java:5186)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getFileInfo(FSNamesystem.java:1994)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.getFileInfo(NameNode.java:819)
>>         at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
>>         at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>>         at java.lang.reflect.Method.invoke(Method.java:597)
>>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:557)
>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1434)
>>         at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1430)
>>         at java.security.AccessController.doPrivileged(Native Method)
>>         at javax.security.auth.Subject.doAs(Subject.java:396)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1177)
>>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1428)
>>
>>
>> --
>> Thanks,
>> sandeep
>>
>>
>


-- 
Thanks,
sandeep

Reply via email to