Hi Rahul, I think you are treating the svn directory as HIVE_HOME. If you do "ant package", HIVE_HOME should set to build/dist.
Zheng On Tue, Oct 27, 2009 at 1:19 AM, Rahul Pal <[email protected]> wrote: > I copied the files (*hadoop-0.19.0.tar.gz and hadoop-0.20.0.tar.gz*) to * > ~/.ant/cache/hadoop/core/**sources *directory. * *This time ant script > didn't try to download the files, it unzipped the files at * > $HIVE_HOME/build/hadoopcore* and the build was successfull. > > But when it i tried to run the hive through $HIVE_HOME/bin/hive it gave > error message (*-bash: bin/hive: Permission denied*). After changing the > permission through chmod, it says "*Missing Hive Execution Jar: > $HIVE_HOME/lib/hive_exec.jar*". > > After copying the jar(hive_exec.jar) through *cp > $HIVE_HOME/build/dist/lib/hive_exec.jar > lib/ *to lib, it says *Missing Hive MetaStore Jar* > > Also does i need to change the configs in those unzipped hadoop > directory(hadoop-0.20.0, hadoop-0.19.0), when i already have pre-configured > hadoop running at different path. > > Please provide inputs, is there any straight forward way of integrating > hive with already present Hadoop. > > Thanks > Rahul > > > > On Mon, Oct 26, 2009 at 9:11 PM, Matt Pestritto <[email protected]>wrote: > >> This also came up in a thread last week. >> >> Same thing happened to me. >> >> My temp workaround was: >> cd ~/.ant/cache/hadoop/core/sources >> wget >> http://archive.apache.org/dist/hadoop/core/hadoop-0.20.1/hadoop-0.20.1.tar.gz >> >> Then rebuild hive. Ivy will not try to download the source again. >> >> Thanks >> -Matt >> >> >> On Mon, Oct 26, 2009 at 11:09 AM, Edward Capriolo >> <[email protected]>wrote: >> >>> On Mon, Oct 26, 2009 at 3:27 AM, Carl Steinbach <[email protected]> >>> wrote: >>> > Hi Rahul, >>> > >>> > Please follow these steps: >>> > >>> > 1) In your hive source directory run 'ant clean'. >>> > 2) remove the contents of ~/.ant/cache/hadoop/core/sources >>> > 3) Download the following files to ~/.ant/cache/hadoop/core/sources: >>> > hadoop-0.17.2.1.tar.gz >>> > hadoop-0.17.2.1.tar.gz.asc >>> > hadoop-0.18.3.tar.gz >>> > hadoop-0.18.3.tar.gz.asc >>> > hadoop-0.19.0.tar.gz >>> > hadoop-0.19.0.tar.gz.asc >>> > hadoop-0.20.0.tar.gz >>> > hadoop-0.20.0.tar.gz.asc >>> > >>> > 4) For each hadoop-xxx.tar.gz file, compute the sha1 checksum using >>> sha1sum, >>> > and verify that it matches the sha1 checksum in the corresponding .asc >>> file. >>> > >>> > If it does not match then the file is corrupt and you need to try >>> > downloading it again. >>> > >>> > 5) Try building Hive again following the instructions on the wiki. You >>> > shouldn't have any problems if you verified the checksums. >>> > >>> > As an additional note, if you don't care about support for Hadoop >>> 0.17.2.1, >>> > or 0.18, etc, you can disable support for these versions (and skip the >>> > download) by removing the references to these versions that >>> shims/build.xml >>> > and shims/ivy.xml >>> > >>> >> >>> >> Also I want to use hive on top of current working hadoop cluster. >>> >> Please provide some inputs. >>> > >>> > You need to set HADOOP_HOME and make sure that HADOOP_HOME/conf (or >>> > HADOOP_CONF_DIR if you have this variable set) contains valid >>> configuration >>> > files for your current working hadoop cluster. See the following page >>> for >>> > more information: >>> http://wiki.apache.org/hadoop/GettingStartedWithHadoop >>> > >>> > Thanks. >>> > >>> > Carl >>> > >>> > On Sun, Oct 25, 2009 at 11:13 PM, Rahul Pal <[email protected]> >>> wrote: >>> >> >>> >> Its still not working Carl. >>> >> First it tries to download the hadoop-0.17.2.1.tar.gz and then try the >>> >> upper versions of hadoop. >>> >> Somehow build script unable to download the hadoop packages and shows >>> >> mismatch in size >>> >> error ( Downloaded file size doesn't match expected Content Length for >>> >> >>> http://archive.apache.org/dist/hadoop/core/hadoop-0.17.2.1/hadoop-0.17.2.1.tar.gz. >>> ). >>> > >>> > >>> >> >>> >> Also I want to use hive on top of current working hadoop cluster. >>> >> Please provide some inputs. >>> > >>> > >>> >> >>> >> Thanks & Regards >>> >> Rahul >>> >> >>> >> >>> >> On Fri, Oct 23, 2009 at 1:29 PM, Carl Steinbach <[email protected]> >>> wrote: >>> >>> >>> >>> Hi Rahul, >>> >>> >>> >>> One solution is to manually download the files to >>> >>> ~/.ant/cache/hadoop/core/sources/ >>> >>> >>> >>> This should prevent ivy from trying (and failing) to download them >>> >>> itself. >>> >>> >>> >>> Carl >>> >>> >>> >>> On Thu, Oct 22, 2009 at 10:41 PM, Rahul Pal <[email protected]> >>> wrote: >>> >>>> >>> >>>> Hi guyz, >>> >>>> I am trying to build Hive from >>> >>>> the trunk - not sure whether I'll be able to do it or not - because >>> >>>> every time I tried that, the build process started downloading all >>> >>>> versions of hadoop and failed with an error saying that the >>> downloaded >>> >>>> size didn't match the expected size... >>> >>>> Please provide some input. >>> >>>> >>> >>>> Thanks & Regards >>> >>>> Rahul Pal >>> >>>> Software Engineer - Discover >>> >>>> One97 Communications (P) Ltd >>> >>>> B121, Sector 5, Noida, UP 201301 >>> >>>> >>> >>>> P: + 91 120 4770770 Extn:312 >>> >>>> M: + 91 9873005998 >>> >>>> W: www.one97world.com >>> >>>> >>> >>>> >>> >>> >>> >> >>> > >>> > >>> >>> All, >>> >>> On a related note this is the second person with this checksum issue. >>> I am not seeing the problem myself but I have not updated from trunk >>> recently. Does anyone know what the fix, not the work around is? >>> >>> Edward >>> >> >> > -- Yours, Zheng
