*rpool/tmp 10G 10G 0K 100% /tmp* * * This might be the source of your problem as I mentioned earlier. Try freeing some space here and then try again.
On Fri, Jul 20, 2012 at 11:34 AM, comptech geeky <comptechge...@gmail.com>wrote: > After trying "df -kh". I got below result. > > *bash-3.00$ df -kh* > *Filesystem size used avail capacity Mounted on* > *rpool/ROOT/sol10 916G 30G 668G 5% /* > */devices 0K 0K 0K 0% /devices* > *ctfs 0K 0K 0K 0% /system/contract* > *proc 0K 0K 0K 0% /proc* > *mnttab 0K 0K 0K 0% /etc/mnttab* > *swap 31G 656K 31G 1% /etc/svc/volatile* > *objfs 0K 0K 0K 0% /system/object* > *sharefs 0K 0K 0K 0% /etc/dfs/sharetab* > */usr/lib/libc/libc_hwcap2.so.1* > * 698G 30G 668G 5% /lib/libc.so.1* > *fd 0K 0K 0K 0% /dev/fd* > *rpool/ROOT/sol10/var 20G 10G 9.7G 52% /var* > *rpool/tmp 10G 10G 0K 100% /tmp* > *swap 31G 20K 31G 1% /var/run* > *lvsaishdc3in0001data/data* > * 32T 27T 2.4T 92% /data* > *lvsaishdc3in0001data/data/b_apdpds* > * 1.0T 8.5G 1016G 1% /data/b_apdpds* > *lvsaishdc3in0001data/data/b_bids* > * 100G 75G 25G 76% /data/b_bids* > *lvsaishdc3in0001data/data/b_sbe* > * 100G 51K 100G 1% /data/b_sbe* > *lvsaishdc3in0001data/data/b_selling* > * 500G 298G 202G 60% /data/b_selling* > *lvsaishdc3in0001data/data/imk* > * 3.0T 2.7T 293G 91% /data/inbound/sq/imk* > *rpool/export 916G 23K 668G 1% /export* > *rpool/export/home 175G 118G 57G 68% /export/home* > *rpool 916G 34K 668G 1% /rpool* > * > * > > > On Fri, Jul 20, 2012 at 7:42 AM, kulkarni.swar...@gmail.com < > kulkarni.swar...@gmail.com> wrote: > >> Seems to me like you might be just running out of disk space on one of >> the partitions. What does the output of "df -kh" say? >> >> Also, I just speculate that it might be your "/tmp" directory out of >> space because that is where hive tries to dump a bunch of log entries >> before it starts up. (/tmp/<user>/hive.log). >> >> >> On Fri, Jul 20, 2012 at 3:12 AM, comptech geeky >> <comptechge...@gmail.com>wrote: >> >>> Whenever I am typing Hive at the command prompt, I am getting the below >>> exception. What does that mean? >>> * >>> * >>> *$ bash* >>> *bash-3.00$ hive* >>> *Exception in thread "main" java.io.IOException: Disc quota exceeded* >>> * at java.io.UnixFileSystem.createFileExclusively(Native Method)* >>> * at java.io.File.checkAndCreate(File.java:1704)* >>> * at java.io.File.createTempFile(File.java:1792)* >>> * at org.apache.hadoop.util.RunJar.main(RunJar.java:115)* >>> *bash-3.00$* >>> >>> Any suggestions why is it happening? >>> >> >> >> >> -- >> Swarnim >> > > -- Swarnim