Hi Nitin

Thanks for your reply.

I am using sqoop *1.4.1-incubating* version. In the sqoop releases download
page the is no such version you are referring. Please correct me if I am
wrong.

Delete the warehouse folder and import is working fine, but my tables
having GB of data, so every time delete & import is not a good answer to my
solution. I am working on a solution to our production system.

Is there any way to solve this issue.

Thanks.

On Fri, Apr 13, 2012 at 11:13 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:

> Hi Roshan,
>
> I guess you are using sqoop version older than 17.
>
> You are facing similar issue mentioned in 
> SQOOP-216<https://issues.cloudera.org/browse/SQOOP-216>
>
> You can try to delete the directory already existing.
>
> Thanks,
> Nitin
>
>
> On Fri, Apr 13, 2012 at 6:12 PM, Roshan Pradeep <codeva...@gmail.com>wrote:
>
>> Hadoop - 0.20.2
>> Hive - 0.8.1
>>
>> Thanks.
>>
>>
>> On Fri, Apr 13, 2012 at 5:03 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:
>>
>>> can you tell us what is
>>> 1) hive version
>>> 2) hadoop version that you are using?
>>>
>>>
>>>
>>>
>>>
>>> On Fri, Apr 13, 2012 at 12:23 PM, Roshan Pradeep <codeva...@gmail.com>wrote:
>>>
>>>> Hi
>>>>
>>>> I want to import the updated data from my source (PostgreSQL) to hive
>>>> based on a column (lastmodifiedtime) in postgreSQL
>>>>
>>>> *The command I am using*
>>>>
>>>> /app/sqoop/bin/sqoop import --hive-table users --connect
>>>> jdbc:postgresql:/<server_url>/<database> --table users --username XXXXXXX
>>>> --password YYYYYY --hive-home /app/hive --hive-import --incremental
>>>> lastmodified --check-column lastmodifiedtime
>>>>
>>>> *With the above command, I am getting the below error*
>>>>
>>>> 12/04/13 16:31:21 INFO orm.CompilationManager: Writing jar file:
>>>> /tmp/sqoop-root/compile/11ce8600a5656ed49e631a260c387692/users.jar
>>>> 12/04/13 16:31:21 INFO tool.ImportTool: Incremental import based on
>>>> column "lastmodifiedtime"
>>>> 12/04/13 16:31:21 INFO tool.ImportTool: Upper bound value: '2012-04-13
>>>> 16:31:21.865429'
>>>> 12/04/13 16:31:21 WARN manager.PostgresqlManager: It looks like you are
>>>> importing from postgresql.
>>>> 12/04/13 16:31:21 WARN manager.PostgresqlManager: This transfer can be
>>>> faster! Use the --direct
>>>> 12/04/13 16:31:21 WARN manager.PostgresqlManager: option to exercise a
>>>> postgresql-specific fast path.
>>>> 12/04/13 16:31:21 INFO mapreduce.ImportJobBase: Beginning import of
>>>> users
>>>> 12/04/13 16:31:23 ERROR tool.ImportTool: Encountered IOException
>>>> running import job: org.apache.hadoop.mapred.FileAlreadyExistsException:
>>>> Output directory users already exists
>>>>         at
>>>> org.apache.hadoop.mapreduce.lib.output.FileOutputFormat.checkOutputSpecs(FileOutputFormat.java:123)
>>>>         at
>>>> org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:770)
>>>>         at org.apache.hadoop.mapreduce.Job.submit(Job.java:432)
>>>>         at
>>>> org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:447)
>>>>         at
>>>> org.apache.sqoop.mapreduce.ImportJobBase.runJob(ImportJobBase.java:141)
>>>>         at
>>>> org.apache.sqoop.mapreduce.ImportJobBase.runImport(ImportJobBase.java:201)
>>>>         at
>>>> org.apache.sqoop.manager.SqlManager.importTable(SqlManager.java:413)
>>>>         at
>>>> org.apache.sqoop.manager.PostgresqlManager.importTable(PostgresqlManager.java:102)
>>>>         at
>>>> org.apache.sqoop.tool.ImportTool.importTable(ImportTool.java:380)
>>>>         at org.apache.sqoop.tool.ImportTool.run(ImportTool.java:453)
>>>>         at org.apache.sqoop.Sqoop.run(Sqoop.java:145)
>>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>>>>         at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:181)
>>>>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:220)
>>>>         at org.apache.sqoop.Sqoop.runTool(Sqoop.java:229)
>>>>         at org.apache.sqoop.Sqoop.main(Sqoop.java:238)
>>>>         at com.cloudera.sqoop.Sqoop.main(Sqoop.java:57)
>>>>
>>>> According to the above, it identify the updated data from postgreSQL,
>>>> but it says output directory already exists. Could someone please help me
>>>> to correct this issue.
>>>>
>>>> Thanks.
>>>>
>>>
>>>
>>>
>>> --
>>> Nitin Pawar
>>>
>>>
>>
>
>
> --
> Nitin Pawar
>
>

Reply via email to