> The 1600MB number looks like record-level compression. Are you sure
> you've turned on block compression?


Here's the exact snippet from my shell script. Do I have to set these
configuration parameters directly in the hadoop configuration file:

    ${HIVE_COMMAND} -e "set hive.exec.compress.output=true; set
io.seqfile.compression.type=BLOCK; set
mapred.output.compression.codec=org.apache.hadoop.io.compress.GzipCodec; set
io.seqfile.compress.blocksize=50000000; insert overwrite table
raw_compressed partition(dt='${D}') select line from raw where dt='${D}'"

Saurabh.
-- 
http://nandz.blogspot.com
http://foodieforlife.blogspot.com

Reply via email to