Hi, The default HWI (hive web interface) provides some basic metadata, but don't think file sizes are included. In any case, you can query using the common hadoop dfs commands. The default warehouse directory is as set in your hive conf xml.
Amogh On 7/26/10 2:30 PM, "vaibhav negi" <[email protected]> wrote: Hi, Thanks amogh. How can i browse actual physical location of hive tables juts like i see mysql tables in mysql directory. I want to check actual disk space consumed by hive tables. Vaibhav Negi On Mon, Jul 26, 2010 at 1:55 PM, Amogh Vasekar <[email protected]> wrote: Hi, You can create an external table pointing to data already on hdfs and specifying the delimiter- CREATE EXTERNAL TABLE page_view_stg(viewTime INT, userid BIGINT, page_url STRING, referrer_url STRING, ip STRING COMMENT 'IP Address of the User', country STRING COMMENT 'country of origination') COMMENT 'This is the staging page view table' ROW FORMAT DELIMITED FIELDS TERMINATED BY '44' LINES TERMINATED BY '12' STORED AS TEXTFILE LOCATION '/user/data/staging/page_view'; http://wiki.apache.org/hadoop/Hive/Tutorial#Creating_Tables for more HTH, Amogh On 7/26/10 1:02 PM, "vaibhav negi" <[email protected] <http://[email protected]> > wrote: Hi, Is there some way to load csv file into hive? Vaibhav Negi
