Hi, Hadoop -dfs command show logical path /user/hive/warehouse. How can i see where this directory exists physically ?
Vaibhav Negi On Mon, Jul 26, 2010 at 2:45 PM, Amogh Vasekar <[email protected]> wrote: > Hi, > The default HWI (hive web interface) provides some basic metadata, but > don’t think file sizes are included. In any case, you can query using the > common hadoop dfs commands. The default warehouse directory is as set in > your hive conf xml. > > Amogh > > > > On 7/26/10 2:30 PM, "vaibhav negi" <[email protected]> wrote: > > Hi, > > Thanks amogh. > How can i browse actual physical location of hive tables juts like i see > mysql tables in mysql directory. I want to check actual disk space consumed > by hive tables. > > > > Vaibhav Negi > > > On Mon, Jul 26, 2010 at 1:55 PM, Amogh Vasekar <[email protected]> > wrote: > > Hi, > You can create an external table pointing to data already on hdfs and > specifying the delimiter- > CREATE EXTERNAL TABLE page_view_stg(viewTime INT, userid BIGINT, > page_url STRING, referrer_url STRING, > ip STRING COMMENT 'IP Address of the User', > country STRING COMMENT 'country of origination') > COMMENT 'This is the staging page view table' > ROW FORMAT DELIMITED FIELDS TERMINATED BY '44' LINES TERMINATED BY '12' > STORED AS TEXTFILE > LOCATION '/user/data/staging/page_view'; > > http://wiki.apache.org/hadoop/Hive/Tutorial#Creating_Tables for more > > HTH, > Amogh > > > > On 7/26/10 1:02 PM, "vaibhav negi" <[email protected] < > http://[email protected]> > wrote: > > Hi, > > Is there some way to load csv file into hive? > > Vaibhav Negi > > > >
