any solution guys,badly stuck in this [?][?][?] On Tue, Sep 4, 2012 at 4:28 PM, Visioner Sadak <[email protected]>wrote:
> Thanks bejoy, actually my hadoop is also on windows(i have installed it in > psuedo-distributed mode for testing) its not a remote cluster.... > > > On Tue, Sep 4, 2012 at 3:38 PM, Bejoy KS <[email protected]> wrote: > >> ** >> Hi >> >> You are running tomact on a windows machine and trying to connect to a >> remote hadoop cluster from there. Your core site has >> >> <name> >> fs.default.name</name> >> <value>hdfs://localhost:9000</value> >> >> But It is localhost here.( I assume you are not running hadoop on this >> windows environment for some testing) >> >> You need to have the exact configuration files and hadoop jars from the >> cluster machines on this tomcat environment as well. I mean on the >> classpath of your application. >> Regards >> Bejoy KS >> >> Sent from handheld, please excuse typos. >> ------------------------------ >> *From: *Visioner Sadak <[email protected]> >> *Date: *Tue, 4 Sep 2012 15:31:25 +0530 >> *To: *<[email protected]> >> *ReplyTo: *[email protected] >> *Subject: *Re: Integrating hadoop with java UI application deployed on >> tomcat >> >> also getting one more error >> >> * >> >> org.apache.hadoop.ipc.RemoteException >> *: Server IPC version 5 cannot communicate with client version 4 >> >> On Tue, Sep 4, 2012 at 2:44 PM, Visioner Sadak >> <[email protected]>wrote: >> >>> Thanks shobha tried adding conf folder to tomcats classpath still >>> getting same error >>> >>> >>> Call to localhost/127.0.0.1:9000 failed on local exception: >>> java.io.IOException: An established connection was aborted by the software >>> in your host machine >>> >>> On Tue, Sep 4, 2012 at 11:18 AM, Mahadevappa, Shobha < >>> [email protected]> wrote: >>> >>>> Hi,**** >>>> >>>> Try adding the hadoop/conf directory in the TOMCAT’s classpath **** >>>> >>>> ** ** >>>> >>>> Ex : >>>> CLASSPATH=/usr/local/Apps/hbase-0.90.4/conf:/usr/local/Apps/hadoop-0.20.203.0/conf: >>>> **** >>>> >>>> ** ** >>>> >>>> ** ** >>>> >>>> ** ** >>>> >>>> Regards,**** >>>> >>>> *Shobha M ***** >>>> >>>> ** ** >>>> >>>> *From:* Visioner Sadak [mailto:[email protected]] >>>> *Sent:* 03 September 2012 PM 04:01 >>>> *To:* [email protected] >>>> >>>> *Subject:* Re: Integrating hadoop with java UI application deployed on >>>> tomcat**** >>>> >>>> ** ** >>>> >>>> Thanks steve thers nothing in logs and no exceptions as well i found >>>> that some file is created in my F:\user with directory name but its not >>>> visible inside my hadoop browse filesystem directories i also added the >>>> config by using the below method **** >>>> >>>> hadoopConf.addResource(**** >>>> >>>> "F:/hadoop-0.22.0/conf/core-site.xml"); **** >>>> >>>> when running thru WAR printing out the filesystem i m getting >>>> org.apache.hadoop.fs.LocalFileSystem@9cd8db **** >>>> >>>> when running an independet jar within hadoop i m getting >>>> DFS[DFSClient[clientName=DFSClient_296231340, ugi=dell]]**** >>>> >>>> when running an independet jar i m able to do uploads....**** >>>> >>>> **** >>>> >>>> just wanted to know will i have to add something in my classpath of >>>> tomcat or is there any other configurations of core-site.xml that i am >>>> missing out..thanks for your help.....**** >>>> >>>> **** >>>> >>>> ** ** >>>> >>>> On Sat, Sep 1, 2012 at 1:38 PM, Steve Loughran <[email protected]> >>>> wrote:**** >>>> >>>> ** ** >>>> >>>> well, it's worked for me in the past outside Hadoop itself:**** >>>> >>>> ** ** >>>> >>>> >>>> http://smartfrog.svn.sourceforge.net/viewvc/smartfrog/trunk/core/hadoop-components/hadoop-ops/src/org/smartfrog/services/hadoop/operations/utils/DfsUtils.java?revision=8882&view=markup >>>> **** >>>> >>>> ** ** >>>> >>>> 1. Turn logging up to DEBUG**** >>>> 2. Make sure that the filesystem you've just loaded is what you >>>> expect, by logging its value. It may turn out to be file:///, >>>> because the normal Hadoop site-config.xml isn't being picked up**** >>>> >>>> **** >>>> >>>> ** ** >>>> >>>> On Fri, Aug 31, 2012 at 1:08 AM, Visioner Sadak < >>>> [email protected]> wrote:**** >>>> >>>> but the problem is that my code gets executed with the warning but >>>> file is not copied to hdfs , actually i m trying to copy a file from local >>>> to hdfs **** >>>> >>>> **** >>>> >>>> Configuration hadoopConf=new Configuration(); >>>> //get the default associated file system >>>> FileSystem fileSystem=FileSystem.get(hadoopConf); >>>> // HarFileSystem harFileSystem= new HarFileSystem(fileSystem); >>>> //copy from lfs to hdfs >>>> fileSystem.copyFromLocalFile(new Path("E:/test/GANI.jpg"),new >>>> Path("/user/TestDir/")); **** >>>> >>>> **** >>>> >>>> ** ** >>>> >>>> ** ** >>>> >>>> >>>> ______________________________________________________________________ >>>> Disclaimer:This email and any attachments are sent in strictest >>>> confidence for the sole use of the addressee and may contain legally >>>> privileged, confidential, and proprietary data. If you are not the intended >>>> recipient, please advise the sender by replying promptly to this email and >>>> then delete and destroy this email and any attachments without any further >>>> use, copying or forwarding >>>> >>> >>> >> >
<<33A.gif>>
