Hi,
The default HWI (hive web interface) provides some basic metadata, but don't 
think file sizes are included. In any case, you can query using the common 
hadoop dfs commands. The default warehouse directory is as set in your hive 
conf xml.

Amogh


On 7/26/10 2:30 PM, "vaibhav negi" <[email protected]> wrote:

Hi,

Thanks amogh.
How can i browse actual physical location  of hive tables juts like i see mysql 
tables in mysql directory. I want to check actual disk space consumed by hive 
tables.



Vaibhav Negi


On Mon, Jul 26, 2010 at 1:55 PM, Amogh Vasekar <[email protected]> wrote:
Hi,
You can create an external table pointing to data already on hdfs and 
specifying the delimiter-
CREATE EXTERNAL TABLE page_view_stg(viewTime INT, userid BIGINT,
                    page_url STRING, referrer_url STRING,
                    ip STRING COMMENT 'IP Address of the User',
                    country STRING COMMENT 'country of origination')
    COMMENT 'This is the staging page view table'
    ROW FORMAT DELIMITED FIELDS TERMINATED BY '44' LINES TERMINATED BY '12'
    STORED AS TEXTFILE
    LOCATION '/user/data/staging/page_view';

http://wiki.apache.org/hadoop/Hive/Tutorial#Creating_Tables   for more

HTH,
Amogh



On 7/26/10 1:02 PM, "vaibhav negi" <[email protected] 
<http://[email protected]> > wrote:

Hi,

Is there some way to load csv file into hive?

Vaibhav Negi



Reply via email to