[jira] [Commented] (SQOOP-3022) sqoop export for Oracle generates tremendous amounts of redo logs

2016-10-12 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-3022?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15570633#comment-15570633
 ] 

David Robson commented on SQOOP-3022:
-

I haven't tested all the various scenarios in a while but basically if you use 
partitioning by setting "oraoop.partitioned" to "TRUE" it should create a table 
per mapper and do direct path inserts. 
http://sqoop.apache.org/docs/1.4.6/SqoopUserGuide.html#_partitioning

If you don't want to use partitioning but want to disable redo logs then you 
can set "oraoop.oracle.append.values.hint.usage" to "ON" which will use the 
APPEND_VALUES hint for the insert statements into the target table. 
http://sqoop.apache.org/docs/1.4.6/SqoopUserGuide.html#_oraoop_oracle_append_values_hint_usage

So I think it already does what you require so we should be able to close this 
issue?

> sqoop export for Oracle generates tremendous amounts of redo logs
> -
>
> Key: SQOOP-3022
> URL: https://issues.apache.org/jira/browse/SQOOP-3022
> Project: Sqoop
>  Issue Type: Bug
>  Components: codegen, connectors, connectors/oracle
>Affects Versions: 1.4.3, 1.4.4, 1.4.5, 1.4.6
>Reporter: Ruslan Dautkhanov
>  Labels: export, oracle
>
> Sqoop export for Oracle generates tremendous amounts of redo logs (comparable 
> to export size or more).
> We have put target tables in nologgin mode, but Oracle will still generate 
> redo logs unless +APPEND Oracle insert hint is used.
> See https://oracle-base.com/articles/misc/append-hint for examples.
> Please add an option for sqoop to generate insert statements in Oracle with 
> APPEND statement. Our databases are swamped with redo/archived logs whenever 
> we sqoop data to them. This is easily avoidable. And from business 
> prospective sqooping to staging tables in nologgin mode is totally fine.
> Thank you.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (SQOOP-2737) Cannot import table from Oracle with column with spaces in name

2015-12-10 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-2737?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15051886#comment-15051886
 ] 

David Robson commented on SQOOP-2737:
-

Can you please try the patch for 
https://issues.apache.org/jira/browse/SQOOP-2723

I have fixed the code to quote the Oracle columns so it should also resolve the 
issue with spaces. This is for the direct connector (using --direct option). 
You'll need to build the current branch. Or if you need help I can do this for 
you if you let me know what version of Hadoop you are running?

> Cannot import table from Oracle with column with spaces in name
> ---
>
> Key: SQOOP-2737
> URL: https://issues.apache.org/jira/browse/SQOOP-2737
> Project: Sqoop
>  Issue Type: Bug
>  Components: connectors/oracle
>Affects Versions: 1.4.5
>Reporter: Jakub HavlĂ­k
>Priority: Critical
>
> When importing database containing table(s) with spaces in column name(s) 
> following error is displayed:
> {code}
> Error: java.io.IOException: SQLException in nextKeyValue
>   at 
> org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:277)
>   at 
> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:553)
>   at 
> org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue(MapContextImpl.java:80)
>   at 
> org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.nextKeyValue(WrappedMapper.java:91)
>   at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
>   at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:784)
>   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
>   at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
>   at java.security.AccessController.doPrivileged(Native Method)
>   at javax.security.auth.Subject.doAs(Subject.java:422)
>   at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
>   at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: java.sql.SQLSyntaxErrorException: ORA-00923: FROM keyword not 
> found where expected
>   at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:447)
>   at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:396)
>   at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:951)
>   at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:513)
>   at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:227)
>   at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:531)
>   at 
> oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:208)
>   at 
> oracle.jdbc.driver.T4CPreparedStatement.executeForDescribe(T4CPreparedStatement.java:886)
>   at 
> oracle.jdbc.driver.OracleStatement.executeMaybeDescribe(OracleStatement.java:1175)
>   at 
> oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1296)
>   at 
> oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3613)
>   at 
> oracle.jdbc.driver.OraclePreparedStatement.executeQuery(OraclePreparedStatement.java:3657)
>   at 
> oracle.jdbc.driver.OraclePreparedStatementWrapper.executeQuery(OraclePreparedStatementWrapper.java:1495)
>   at 
> org.apache.sqoop.mapreduce.db.DBRecordReader.executeQuery(DBRecordReader.java:111)
>   at 
> org.apache.sqoop.mapreduce.db.DBRecordReader.nextKeyValue(DBRecordReader.java:235)
>   ... 11 more
> {code}
> As a enterprise user I would expect that I can import tables with spaces in a 
> column name with no issues.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (SQOOP-2729) Sqoop2: Integration Tests: In ShortStories.java we have a varchar(10000) column, which is too long for oracle

2015-12-08 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-2729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15047703#comment-15047703
 ] 

David Robson commented on SQOOP-2729:
-

FYI you can work around this - but only if running Oracle 12c. So I think not 
using more than 4000 chars is the easier solution.

https://docs.oracle.com/database/121/SQLRF/sql_elements001.htm#SQLRF55623

> Sqoop2: Integration Tests: In ShortStories.java we have a varchar(1) 
> column, which is too long for oracle
> -
>
> Key: SQOOP-2729
> URL: https://issues.apache.org/jira/browse/SQOOP-2729
> Project: Sqoop
>  Issue Type: Bug
>Affects Versions: 1.99.6
>Reporter: Abraham Fine
>Assignee: Abraham Fine
> Attachments: SQOOP-2729.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (SQOOP-2723) Oracle connector not working with lowercase columns

2015-12-08 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-2723?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15047628#comment-15047628
 ] 

David Robson commented on SQOOP-2723:
-

Thanks Jarcec. I had put it on review board too but forgot to link it!

I think those failures are unrelated to anything I have changed?

> Oracle connector not working with lowercase columns
> ---
>
> Key: SQOOP-2723
> URL: https://issues.apache.org/jira/browse/SQOOP-2723
> Project: Sqoop
>  Issue Type: Bug
>  Components: connectors/oracle
>Affects Versions: 1.4.6
>Reporter: David Robson
>Assignee: David Robson
> Fix For: 1.4.7
>
> Attachments: SQOOP-2723.patch
>
>
> For example the following table does not work for imports or exports:
> CREATE TABLE "TestV500"."Xy" 
>(  "a" NVARCHAR2(255) DEFAULT NULL, 
>   "x" NUMBER(10,0) DEFAULT NULL, 
>   "y" NUMBER(10,0) DEFAULT NULL
>);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (SQOOP-2723) Oracle connector not working with lowercase columns

2015-12-07 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-2723?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-2723:

Attachment: SQOOP-2723.patch

> Oracle connector not working with lowercase columns
> ---
>
> Key: SQOOP-2723
> URL: https://issues.apache.org/jira/browse/SQOOP-2723
> Project: Sqoop
>  Issue Type: Bug
>  Components: connectors/oracle
>Affects Versions: 1.4.6
>Reporter: David Robson
>Assignee: David Robson
> Fix For: 1.4.7
>
> Attachments: SQOOP-2723.patch
>
>
> For example the following table does not work for imports or exports:
> CREATE TABLE "TestV500"."Xy" 
>(  "a" NVARCHAR2(255) DEFAULT NULL, 
>   "x" NUMBER(10,0) DEFAULT NULL, 
>   "y" NUMBER(10,0) DEFAULT NULL
>);



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-2723) Oracle connector not working with lowercase columns

2015-12-07 Thread David Robson (JIRA)
David Robson created SQOOP-2723:
---

 Summary: Oracle connector not working with lowercase columns
 Key: SQOOP-2723
 URL: https://issues.apache.org/jira/browse/SQOOP-2723
 Project: Sqoop
  Issue Type: Bug
  Components: connectors/oracle
Affects Versions: 1.4.6
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.7


For example the following table does not work for imports or exports:

CREATE TABLE "TestV500"."Xy" 
   ("a" NVARCHAR2(255) DEFAULT NULL, 
"x" NUMBER(10,0) DEFAULT NULL, 
"y" NUMBER(10,0) DEFAULT NULL
   );




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-2711) Exit command no longer works in sqoop2-shell

2015-11-23 Thread David Robson (JIRA)
David Robson created SQOOP-2711:
---

 Summary: Exit command no longer works in sqoop2-shell
 Key: SQOOP-2711
 URL: https://issues.apache.org/jira/browse/SQOOP-2711
 Project: Sqoop
  Issue Type: Bug
  Components: sqoop2-shell
Affects Versions: 1.99.7
Reporter: David Robson


A recent change seems to have broken the exit command. Ctrl-D still works as 
normal.

sqoop:000> exit
Unknown property: exit




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-2699) Improve Oracle parameters

2015-11-18 Thread David Robson (JIRA)
David Robson created SQOOP-2699:
---

 Summary: Improve Oracle parameters
 Key: SQOOP-2699
 URL: https://issues.apache.org/jira/browse/SQOOP-2699
 Project: Sqoop
  Issue Type: Bug
  Components: connectors/oracle
Affects Versions: 1.99.7
Reporter: David Robson
 Fix For: 1.99.7


Some of the Oracle parameters need to be improved to use Lists, Enums etc. Can 
also improve the names of them.

Also need to add the ability to filter the columns on the Oracle table.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (SQOOP-2699) Improve Oracle parameters

2015-11-18 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-2699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-2699:

Attachment: SQOOP-2699.patch

> Improve Oracle parameters
> -
>
> Key: SQOOP-2699
> URL: https://issues.apache.org/jira/browse/SQOOP-2699
> Project: Sqoop
>  Issue Type: Bug
>  Components: connectors/oracle
>Affects Versions: 1.99.7
>Reporter: David Robson
> Fix For: 1.99.7
>
> Attachments: SQOOP-2699.patch
>
>
> Some of the Oracle parameters need to be improved to use Lists, Enums etc. 
> Can also improve the names of them.
> Also need to add the ability to filter the columns on the Oracle table.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (SQOOP-2595) Add Oracle connector to Sqoop 2

2015-11-04 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-2595?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-2595:

Attachment: SQOOP-2595.patch

> Add Oracle connector to Sqoop 2
> ---
>
> Key: SQOOP-2595
> URL: https://issues.apache.org/jira/browse/SQOOP-2595
> Project: Sqoop
>  Issue Type: Improvement
>  Components: connectors/oracle
>Affects Versions: 1.99.7
>Reporter: David Robson
>Assignee: David Robson
> Attachments: SQOOP-2595.patch
>
>
> Migrate the code for the Oracle direct connector into Sqoop 2.
> Initially this will just be the code from the direct connector in Sqoop 1, 
> but eventually we need to migrate the code from the other Oracle connector as 
> well to cover all the scenarios.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (SQOOP-2595) Add Oracle connector to Sqoop 2

2015-11-04 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-2595?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-2595:

Attachment: SQOOP-2595.patch

> Add Oracle connector to Sqoop 2
> ---
>
> Key: SQOOP-2595
> URL: https://issues.apache.org/jira/browse/SQOOP-2595
> Project: Sqoop
>  Issue Type: Improvement
>  Components: connectors/oracle
>Affects Versions: 1.99.7
>Reporter: David Robson
>Assignee: David Robson
> Attachments: SQOOP-2595.patch, SQOOP-2595.patch
>
>
> Migrate the code for the Oracle direct connector into Sqoop 2.
> Initially this will just be the code from the direct connector in Sqoop 1, 
> but eventually we need to migrate the code from the other Oracle connector as 
> well to cover all the scenarios.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-2595) Add Oracle connector to Sqoop 2

2015-09-29 Thread David Robson (JIRA)
David Robson created SQOOP-2595:
---

 Summary: Add Oracle connector to Sqoop 2
 Key: SQOOP-2595
 URL: https://issues.apache.org/jira/browse/SQOOP-2595
 Project: Sqoop
  Issue Type: Improvement
  Components: connectors/oracle
Affects Versions: 1.99.7
Reporter: David Robson
Assignee: David Robson


Migrate the code for the Oracle direct connector into Sqoop 2.
Initially this will just be the code from the direct connector in Sqoop 1, but 
eventually we need to migrate the code from the other Oracle connector as well 
to cover all the scenarios.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-2464) Initializer object is not reused when calling getSchema

2015-08-06 Thread David Robson (JIRA)
David Robson created SQOOP-2464:
---

 Summary: Initializer object is not reused when calling getSchema
 Key: SQOOP-2464
 URL: https://issues.apache.org/jira/browse/SQOOP-2464
 Project: Sqoop
  Issue Type: Bug
Affects Versions: 1.99.6
Reporter: David Robson


In JobManager there is two methods which are called one after the other - 
initializeConnector and getSchemaForConnector. Both these methods do the 
same thing as the first step - create a new instance of the initializer class.

If the same instance of the initializer was shared it means the class could 
keep resources open (such as a connection to the database) and not have to 
re-establish the connection. This might mean a close method needs to be added 
to the initializers as otherwise the getSchema would need to close any 
resources opened in the initialize call - which might seem a bit confusing.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-2465) Initializer and Destroyer should know how many executors will run

2015-08-06 Thread David Robson (JIRA)
David Robson created SQOOP-2465:
---

 Summary: Initializer and Destroyer should know how many executors 
will run
 Key: SQOOP-2465
 URL: https://issues.apache.org/jira/browse/SQOOP-2465
 Project: Sqoop
  Issue Type: Bug
Affects Versions: 1.99.6
Reporter: David Robson


Looking at a job to load data into Oracle as an example - depending on the way 
the user wants to load data, we may be loading data into temporary tables. For 
maximum performance we need to create a separate temporary table for each 
loader - so when the initializer is running we need to know how many loaders 
will run so we can create these temporary tables. Again when the destroyer is 
run we will need to drop these temporary tables - so it will need to know as 
well.

Another example where we need to know this in the initializer - Oracle 
databases may be real application clusters where there is multiple instances 
across multiple machines. For both FROM and TO jobs we spread the load across 
these instances during the initialization phase - so we need to know how many 
loaders / extractors will run.

In the case of a FROM job we could do this in the partition phase - but there 
is no way to achieve this for a TO job. It seems we could either add the 
information into the initialize phase - or add a new partition phase on the TO 
side that is called after the partition phase on the FROM side. It could take 
the details of the partitioned output and match it up to the other side.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-2466) Loader needs to be able to uniquely identify itself

2015-08-06 Thread David Robson (JIRA)
David Robson created SQOOP-2466:
---

 Summary: Loader needs to be able to uniquely identify itself
 Key: SQOOP-2466
 URL: https://issues.apache.org/jira/browse/SQOOP-2466
 Project: Sqoop
  Issue Type: Bug
Affects Versions: 1.99.6
Reporter: David Robson


Following on from SQOOP-2465 - once the balancing of loaders and instances has 
been determined, and temporary tables have been created - the loader needs to 
know which instance it will connect to and which temporary table to use.

Currently all loaders are called with the same parameters - so there's no way 
to distinguish them from each other. The extractor can use the partition input 
to distinguish itself but there is no such equivalent for loaders. Either the 
loader needs some sort of id, or an equivalent of the partition input so it can 
work out which temporary table to use.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-1764) Numeric Overflow when getting extent map

2014-11-18 Thread David Robson (JIRA)
David Robson created SQOOP-1764:
---

 Summary: Numeric Overflow when getting extent map
 Key: SQOOP-1764
 URL: https://issues.apache.org/jira/browse/SQOOP-1764
 Project: Sqoop
  Issue Type: Bug
  Components: connectors/oracle
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson


User is receiving the following error - looks like the block ID can be bigger 
than an int:

14/11/06 14:17:58 ERROR tool.ImportTool: Encountered IOException running import 
job: java.io.IOException: java.sql.SQLException: Numeric Overflow
at 
com.quest.oraoop.OraOopDataDrivenDBInputFormat.getSplits(OraOopDataDrivenDBInputFormat.java:120)
at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:498)
at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:515)
at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:399)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1295)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1292)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
at org.apache.hadoop.mapreduce.Job.submit(Job.java:1292)
at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1313)
at 
org.apache.sqoop.mapreduce.ImportJobBase.doSubmitJob(ImportJobBase.java:198)
at 
org.apache.sqoop.mapreduce.ImportJobBase.runJob(ImportJobBase.java:171)
at 
org.apache.sqoop.mapreduce.ImportJobBase.runImport(ImportJobBase.java:268)
at org.apache.sqoop.manager.SqlManager.importTable(SqlManager.java:665)
at 
com.quest.oraoop.OraOopConnManager.importTable(OraOopConnManager.java:260)
at org.apache.sqoop.tool.ImportTool.importTable(ImportTool.java:497)
at org.apache.sqoop.tool.ImportTool.run(ImportTool.java:605)
at org.apache.sqoop.Sqoop.run(Sqoop.java:143)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:179)
at org.apache.sqoop.Sqoop.runTool(Sqoop.java:218)
at org.apache.sqoop.Sqoop.runTool(Sqoop.java:227)
at org.apache.sqoop.Sqoop.main(Sqoop.java:236)
Caused by: java.sql.SQLException: Numeric Overflow
at 
oracle.jdbc.driver.NumberCommonAccessor.throwOverflow(NumberCommonAccessor.java:4170)
at 
oracle.jdbc.driver.NumberCommonAccessor.getInt(NumberCommonAccessor.java:119)
at 
oracle.jdbc.driver.GeneratedStatement.getInt(GeneratedStatement.java:217)
at 
oracle.jdbc.driver.GeneratedScrollableResultSet.getInt(GeneratedScrollableResultSet.java:522)
at 
oracle.jdbc.driver.GeneratedResultSet.getInt(GeneratedResultSet.java:1350)
at 
com.quest.oraoop.OraOopOracleQueries.getOracleDataChunksExtent(OraOopOracleQueries.java:271)
at 
com.quest.oraoop.OraOopDataDrivenDBInputFormat.getSplits(OraOopDataDrivenDBInputFormat.java:74)
... 23 more




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (SQOOP-1764) Numeric Overflow when getting extent map

2014-11-18 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-1764?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-1764:

Attachment: SQOOP-1764.patch

 Numeric Overflow when getting extent map
 

 Key: SQOOP-1764
 URL: https://issues.apache.org/jira/browse/SQOOP-1764
 Project: Sqoop
  Issue Type: Bug
  Components: connectors/oracle
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-1764.patch


 User is receiving the following error - looks like the block ID can be bigger 
 than an int:
 14/11/06 14:17:58 ERROR tool.ImportTool: Encountered IOException running 
 import job: java.io.IOException: java.sql.SQLException: Numeric Overflow
 at 
 com.quest.oraoop.OraOopDataDrivenDBInputFormat.getSplits(OraOopDataDrivenDBInputFormat.java:120)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:498)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:515)
 at 
 org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:399)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1295)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1292)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)
 at org.apache.hadoop.mapreduce.Job.submit(Job.java:1292)
 at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1313)
 at 
 org.apache.sqoop.mapreduce.ImportJobBase.doSubmitJob(ImportJobBase.java:198)
 at 
 org.apache.sqoop.mapreduce.ImportJobBase.runJob(ImportJobBase.java:171)
 at 
 org.apache.sqoop.mapreduce.ImportJobBase.runImport(ImportJobBase.java:268)
 at 
 org.apache.sqoop.manager.SqlManager.importTable(SqlManager.java:665)
 at 
 com.quest.oraoop.OraOopConnManager.importTable(OraOopConnManager.java:260)
 at org.apache.sqoop.tool.ImportTool.importTable(ImportTool.java:497)
 at org.apache.sqoop.tool.ImportTool.run(ImportTool.java:605)
 at org.apache.sqoop.Sqoop.run(Sqoop.java:143)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
 at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:179)
 at org.apache.sqoop.Sqoop.runTool(Sqoop.java:218)
 at org.apache.sqoop.Sqoop.runTool(Sqoop.java:227)
 at org.apache.sqoop.Sqoop.main(Sqoop.java:236)
 Caused by: java.sql.SQLException: Numeric Overflow
 at 
 oracle.jdbc.driver.NumberCommonAccessor.throwOverflow(NumberCommonAccessor.java:4170)
 at 
 oracle.jdbc.driver.NumberCommonAccessor.getInt(NumberCommonAccessor.java:119)
 at 
 oracle.jdbc.driver.GeneratedStatement.getInt(GeneratedStatement.java:217)
 at 
 oracle.jdbc.driver.GeneratedScrollableResultSet.getInt(GeneratedScrollableResultSet.java:522)
 at 
 oracle.jdbc.driver.GeneratedResultSet.getInt(GeneratedResultSet.java:1350)
 at 
 com.quest.oraoop.OraOopOracleQueries.getOracleDataChunksExtent(OraOopOracleQueries.java:271)
 at 
 com.quest.oraoop.OraOopDataDrivenDBInputFormat.getSplits(OraOopDataDrivenDBInputFormat.java:74)
 ... 23 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (SQOOP-1663) OraOop test cases are not logging any output

2014-11-04 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-1663?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14197093#comment-14197093
 ] 

David Robson commented on SQOOP-1663:
-

This was just so we didn't see the entire log for every build - it only outputs 
the log if there was a problem. In our case it was going to the console - but 
it's a bit different as Sqoop is already putting them into a log file - so you 
could probably just remove the removeAllAppenders call and the System.out 
call and this should mean it always gets logged.

 OraOop test cases are not logging any output
 

 Key: SQOOP-1663
 URL: https://issues.apache.org/jira/browse/SQOOP-1663
 Project: Sqoop
  Issue Type: Bug
Affects Versions: 1.4.5
Reporter: Jarek Jarcec Cecho
Assignee: Jarek Jarcec Cecho
 Fix For: 1.4.6


 Output log from OraOop test cases is always suspiciously empty on my box. It 
 seems that the class OraOopTestCase is doing [some dark 
 magic|https://github.com/apache/sqoop/blob/trunk/src/test/org/apache/sqoop/manager/oracle/OraOopTestCase.java#L188]
  with appenders that results in the empty log.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (SQOOP-1632) Add support for index organized tables to direct connector

2014-10-28 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-1632?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-1632:

Attachment: SQOOP-1632.patch

 Add support for index organized tables to direct connector
 --

 Key: SQOOP-1632
 URL: https://issues.apache.org/jira/browse/SQOOP-1632
 Project: Sqoop
  Issue Type: Improvement
  Components: connectors/oracle
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-1632.patch, SQOOP-1632.patch


 If an index organized table is partitioned - we can use the new chunk by 
 partition functionality on it - so we should add support for this.
 IOTs that are not partitioned will still not be supported by the direct 
 connector.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-1632) Add support for index organized tables to direct connector

2014-10-27 Thread David Robson (JIRA)
David Robson created SQOOP-1632:
---

 Summary: Add support for index organized tables to direct connector
 Key: SQOOP-1632
 URL: https://issues.apache.org/jira/browse/SQOOP-1632
 Project: Sqoop
  Issue Type: Improvement
  Components: connectors/oracle
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson


If an index organized table is partitioned - we can use the new chunk by 
partition functionality on it - so we should add support for this.
IOTs that are not partitioned will still not be supported by the direct 
connector.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (SQOOP-1433) Oracle direct connector fails with Oracle 12c JDBC driver

2014-09-02 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-1433?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-1433:

Attachment: SQOOP-1433.patch

 Oracle direct connector fails with Oracle 12c JDBC driver
 -

 Key: SQOOP-1433
 URL: https://issues.apache.org/jira/browse/SQOOP-1433
 Project: Sqoop
  Issue Type: Bug
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.6

 Attachments: SQOOP-1433.patch


 If you issue a commit statement while auto commit is turned on with the 
 Oracle 12c JDBC driver it throws an error where the 11g drivers did not.
 This causes Sqoop to fail so this needs to be fixed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (SQOOP-1486) Oracle direct connector should use Sqoop merge syntax

2014-08-28 Thread David Robson (JIRA)
David Robson created SQOOP-1486:
---

 Summary: Oracle direct connector should use Sqoop merge syntax
 Key: SQOOP-1486
 URL: https://issues.apache.org/jira/browse/SQOOP-1486
 Project: Sqoop
  Issue Type: Bug
  Components: connectors/oracle
Affects Versions: 1.4.5
Reporter: David Robson


From [~gwenshap]

when running the following sqoop command with OraOop installed:

sqoop export --table mytable --update-key COL1, COL2, COL3 --export-dir 
/user/hive/warehouse/mytable --connect jdbc:oracle:thin:@(DESCRIPTION = 
(ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = 1.1.1.1)(PORT = 1521)))(SDU 
= 5844)(CONNECT_DATA = (ORACLE_SID = oracle))) --username scott --password 
tiger

The job runs succesfully and reports:
INFO mapreduce.ExportJobBase: Exported 54 records.

However not rows actually appears in the table. It appears that Oraoop loads 
the data into staging tables, but not into the user table.

When using the Oraoop specific syntax, the export succeeds and we see data in 
the user table:

sqoop export -Doraoop.export.merge=true 
-Doraoop.update.key.extra.columns=COL2 --table mytable --export-dir 
/user/hive/warehouse/mytable --connect jdbc:oracle:thin:@(DESCRIPTION = 
(ADDRESS_LIST = (ADDRESS = (PROTOCOL = TCP)(HOST = 1.1.1.1)(PORT = 1521)))(SDU 
= 5844)(CONNECT_DATA = (ORACLE_SID = oracle))) --username scott --password 
tiger --input-fields-terminated-by ',' --update-key COL1

Expected behavior:
Either make sure the usual Sqoop syntax works with Oraoop, or issue a suitable 
error message



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (SQOOP-1433) Oracle direct connector fails with Oracle 12c JDBC driver

2014-08-13 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-1433?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14096384#comment-14096384
 ] 

David Robson commented on SQOOP-1433:
-

Yes looks like the same issue! We can just do the same thing for the direct 
connector.

 Oracle direct connector fails with Oracle 12c JDBC driver
 -

 Key: SQOOP-1433
 URL: https://issues.apache.org/jira/browse/SQOOP-1433
 Project: Sqoop
  Issue Type: Bug
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.6


 If you issue a commit statement while auto commit is turned on with the 
 Oracle 12c JDBC driver it throws an error where the 11g drivers did not.
 This causes Sqoop to fail so this needs to be fixed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (SQOOP-1433) Oracle direct connector fails with Oracle 12c JDBC driver

2014-08-12 Thread David Robson (JIRA)
David Robson created SQOOP-1433:
---

 Summary: Oracle direct connector fails with Oracle 12c JDBC driver
 Key: SQOOP-1433
 URL: https://issues.apache.org/jira/browse/SQOOP-1433
 Project: Sqoop
  Issue Type: Bug
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.6


If you issue a commit statement while auto commit is turned on with the Oracle 
12c JDBC driver it throws an error where the 11g drivers did not.

This causes Sqoop to fail so this needs to be fixed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (SQOOP-1422) Integration tests for Oracle connector fail as not using direct option

2014-08-07 Thread David Robson (JIRA)
David Robson created SQOOP-1422:
---

 Summary: Integration tests for Oracle connector fail as not using 
direct option
 Key: SQOOP-1422
 URL: https://issues.apache.org/jira/browse/SQOOP-1422
 Project: Sqoop
  Issue Type: Bug
  Components: connectors/oracle
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson
Priority: Minor


Made the Oracle connector require the direct flag to activate, but didn't 
update the integration tests with this. So they fail now - need to add the 
direct flag to integration tests.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (SQOOP-1422) Integration tests for Oracle connector fail as not using direct option

2014-08-07 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-1422?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-1422:


Attachment: SQOOP-1422.patch

 Integration tests for Oracle connector fail as not using direct option
 --

 Key: SQOOP-1422
 URL: https://issues.apache.org/jira/browse/SQOOP-1422
 Project: Sqoop
  Issue Type: Bug
  Components: connectors/oracle
Affects Versions: 1.4.5
Reporter: David Robson
Assignee: David Robson
Priority: Minor
 Attachments: SQOOP-1422.patch


 Made the Oracle connector require the direct flag to activate, but didn't 
 update the integration tests with this. So they fail now - need to add the 
 direct flag to integration tests.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (SQOOP-1388) One of Export scenario fails with an error : java.sql.SQLException: Unable to perform an exchange subpartition operation for the table ...

2014-07-20 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-1388?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14068182#comment-14068182
 ] 

David Robson commented on SQOOP-1388:
-

This is caused because the requested number of splits is 4, but 
CombineFileInputFormat is generating 5 splits. OraOop is assuming the requested 
number of splits will be the actual number of splits. Need to confirm if this 
is meant to be the case or not.

 One of Export scenario fails with an error  : java.sql.SQLException: Unable 
 to perform an exchange subpartition operation for the table ...
 -

 Key: SQOOP-1388
 URL: https://issues.apache.org/jira/browse/SQOOP-1388
 Project: Sqoop
  Issue Type: Bug
 Environment:   sqoop-1.4.5-SNAPSHOT.bin__hadoop-2.0.4-alpha.tar.gz  
 on CDH4 (melhadoop 01).
Reporter: Vidya Iyengar
Assignee: David Robson

 Ran Export Scenario 13 from the Guy Ley Mar document on SASDB01.
 These are the steps I followed for CASE 13.
   
 drop table junk;
 drop table junk13;
 create table junk as select * from dba_objects;
 sqoop import --direct --connect jdbc:oracle:thin:@melsasdb01:1521:sasdb01 
 --username oraoop --password oraoop --table junk  --where OBJECT_ID100 
 --target-dir junk_a
 sqoop import --direct --connect jdbc:oracle:thin:@melsasdb01:1521:sasdb01 
 --username oraoop --password oraoop --table junk  --where OBJECT_ID= 100 
 and OBJECT_ID  200 --target-dir junk_b
 sqoop export -D oraoop.drop.table=true -Doraoop.template.table=oraoop.junk 
 -Doraoop.nologging=true -Doraoop.partitioned=true -Doraoop.export.merge=false 
 --direct --connect jdbc:oracle:thin:@melsasdb01:1521:sasdb01 --username 
 oraoop --password oraoop --table junk13 --export-dir junk_a
 sqoop export  -Doraoop.export.merge=false --direct --connect 
 jdbc:oracle:thin:@melsasdb01:1521:sasdb01 --username oraoop --password oraoop 
 --table junk13 --export-dir junk_b
 Note: Failed when i ran the first export.
 On the Hadoop file system, only one file contains data remaining 3 files are 
 with size 0. 
 Removing empty files export worked ok.
 Imported folders are available junk_a_bug , junk_b_bug for debugging.
 Attached file contains the output error.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (SQOOP-1344) Add documentation for Oracle connector

2014-07-13 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-1344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-1344:


Attachment: SQOOP-1344.patch

 Add documentation for Oracle connector
 --

 Key: SQOOP-1344
 URL: https://issues.apache.org/jira/browse/SQOOP-1344
 Project: Sqoop
  Issue Type: Task
  Components: connectors/oracle
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.5

 Attachments: SQOOP-1344.patch


 Now the connector is included in Sqoop we need to add any relevant details 
 from the OraOop documentation into the Sqoop documentation.
 This will be in the connectors section of the documentation as discussed on 
 the mailing list.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (SQOOP-732) Fix checkstyle violations on trunk

2014-07-11 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-732?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14058457#comment-14058457
 ] 

David Robson commented on SQOOP-732:


Richard - if you run the checkstyle target from ant - (ant checkstyle in the 
root directory) - the build will generate a report with violations in 
build/checkstyle-errors.html

Currently on trunk there are 235 violations that I can see.

 Fix checkstyle violations on trunk
 --

 Key: SQOOP-732
 URL: https://issues.apache.org/jira/browse/SQOOP-732
 Project: Sqoop
  Issue Type: Task
Reporter: Jarek Jarcec Cecho
Priority: Trivial
  Labels: newbie
 Attachments: patch.diff


 We have checkstyle violations on trunk that should be fixed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (SQOOP-1357) QA testing of Data Connector for Oracle and Hadoop

2014-07-03 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-1357?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-1357:


Assignee: Vidya Iyengar

 QA testing of Data Connector for Oracle and Hadoop
 --

 Key: SQOOP-1357
 URL: https://issues.apache.org/jira/browse/SQOOP-1357
 Project: Sqoop
  Issue Type: Sub-task
  Components: connectors/oracle
Reporter: David Robson
Assignee: Vidya Iyengar
 Fix For: 1.4.5

   Original Estimate: 48h
  Remaining Estimate: 48h

 As part of donating the data connector - Dell would like to perform QA before 
 it is released.
 This should be started on the 16th July.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (SQOOP-1357) QA testing of Data Connector for Oracle and Hadoop

2014-07-02 Thread David Robson (JIRA)
David Robson created SQOOP-1357:
---

 Summary: QA testing of Data Connector for Oracle and Hadoop
 Key: SQOOP-1357
 URL: https://issues.apache.org/jira/browse/SQOOP-1357
 Project: Sqoop
  Issue Type: Sub-task
  Components: connectors/oracle
Reporter: David Robson
 Fix For: 1.4.5


As part of donating the data connector - Dell would like to perform QA before 
it is released.
This should be started on the 16th July.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (SQOOP-1287) Add high performance Oracle connector into Sqoop

2014-06-29 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-1287?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14047063#comment-14047063
 ] 

David Robson commented on SQOOP-1287:
-

Hey Venkat - I checked them they are the same file - what problems do you mean?
Yes don't really need the extra option - I left it there just in case someone 
tried to set it to disable OraOop out of habit...

 Add high performance Oracle connector into Sqoop
 

 Key: SQOOP-1287
 URL: https://issues.apache.org/jira/browse/SQOOP-1287
 Project: Sqoop
  Issue Type: New Feature
  Components: connectors/oracle
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.5

 Attachments: SQOOP-1287.patch


 Dell Software is contributing an Oracle connector for the Sqoop project. We 
 had a discussion about this on the mailing list 
 (http://markmail.org/thread/42tlyjzg5ip245jq), and also a vote 
 (http://markmail.org/thread/wjxpv6zxeo2ioaqj). The results of the vote are 
 here http://markmail.org/message/366aqe6rqy4yji6j



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (SQOOP-1287) Add high performance Oracle connector into Sqoop

2014-06-29 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-1287?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14047357#comment-14047357
 ] 

David Robson commented on SQOOP-1287:
-

Thanks Venkat - looks good. Yeah I think there must have been some Windows line 
endings in there somehow - I'm not sure how they got there as have only used a 
linux machine for this! Anyway sounds like you sorted it out.

 Add high performance Oracle connector into Sqoop
 

 Key: SQOOP-1287
 URL: https://issues.apache.org/jira/browse/SQOOP-1287
 Project: Sqoop
  Issue Type: New Feature
  Components: connectors/oracle
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.5

 Attachments: SQOOP-1287.patch


 Dell Software is contributing an Oracle connector for the Sqoop project. We 
 had a discussion about this on the mailing list 
 (http://markmail.org/thread/42tlyjzg5ip245jq), and also a vote 
 (http://markmail.org/thread/wjxpv6zxeo2ioaqj). The results of the vote are 
 here http://markmail.org/message/366aqe6rqy4yji6j



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (SQOOP-1344) Add documentation for Oracle connector

2014-06-26 Thread David Robson (JIRA)
David Robson created SQOOP-1344:
---

 Summary: Add documentation for Oracle connector
 Key: SQOOP-1344
 URL: https://issues.apache.org/jira/browse/SQOOP-1344
 Project: Sqoop
  Issue Type: Task
  Components: connectors/oracle
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.5


Now the connector is included in Sqoop we need to add any relevant details from 
the OraOop documentation into the Sqoop documentation.
This will be in the connectors section of the documentation as discussed on the 
mailing list.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (SQOOP-1345) Merge the two Oracle connectors into one

2014-06-26 Thread David Robson (JIRA)
David Robson created SQOOP-1345:
---

 Summary: Merge the two Oracle connectors into one
 Key: SQOOP-1345
 URL: https://issues.apache.org/jira/browse/SQOOP-1345
 Project: Sqoop
  Issue Type: Task
  Components: connectors/oracle
Reporter: David Robson


The donated Oracle connector is currently sitting as a --direct connector, 
and the moment. We should just have one Oracle connector - taking the bits that 
each connector does and combining it into one.

This is a good opportunity to refactor some of it - for example moving some of 
the classes to different packages, improving some methods that are too long etc.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (SQOOP-862) Hbase import fails if there is a row where all columns are null

2013-02-07 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-862?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-862:
---

Attachment: SQOOP-862.patch

 Hbase import fails if there is a row where all columns are null
 ---

 Key: SQOOP-862
 URL: https://issues.apache.org/jira/browse/SQOOP-862
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-862.patch, SQOOP-862.patch


 If you try to import a table where any of the rows contain all null values 
 (except for the primary key), the import fails. For example create the 
 following table in Oracle:
 CREATE TABLE employee(id number primary key, test_number number);
 INSERT INTO employee values(1, 123);
 INSERT INTO employee values(2, null);
 COMMIT;
 Then run an import:
 sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username 
 USERNAME --table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE 
 --column-family tst --hbase-create-table
 The Import fails with:
  java.lang.IllegalArgumentException: No columns to insert
   at 
 org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:400)
 Caused by: java.lang.IllegalArgumentException: No columns to insert
   at org.apache.hadoop.hbase.client.HTable.validatePut(HTable.java:950)
   at org.apache.hadoop.hbase.client.HTable.doPut(HTable.java:766)
   at org.apache.hadoop.hbase.client.HTable.put(HTable.java:752)
   at 
 org.apache.sqoop.hbase.HBasePutProcessor.accept(HBasePutProcessor.java:127)
   at 
 org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:128)
   at 
 org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:1)
   at 
 org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.write(MapTask.java:598)
   at 
 org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(TaskInputOutputContextImpl.java:89)
   at 
 org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.write(WrappedMapper.java:112)
   at 
 org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:38)
   at 
 org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:1)
   at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
   at 
 org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)
   at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:725)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
   at 
 org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:232)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
   at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
   at java.util.concurrent.FutureTask.run(FutureTask.java:138)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
   at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (SQOOP-862) Hbase import fails if there is a row where all columns are null

2013-02-07 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-862?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13573966#comment-13573966
 ] 

David Robson commented on SQOOP-862:


Updated patch with documentation changes.

 Hbase import fails if there is a row where all columns are null
 ---

 Key: SQOOP-862
 URL: https://issues.apache.org/jira/browse/SQOOP-862
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-862.patch, SQOOP-862.patch


 If you try to import a table where any of the rows contain all null values 
 (except for the primary key), the import fails. For example create the 
 following table in Oracle:
 CREATE TABLE employee(id number primary key, test_number number);
 INSERT INTO employee values(1, 123);
 INSERT INTO employee values(2, null);
 COMMIT;
 Then run an import:
 sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username 
 USERNAME --table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE 
 --column-family tst --hbase-create-table
 The Import fails with:
  java.lang.IllegalArgumentException: No columns to insert
   at 
 org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:400)
 Caused by: java.lang.IllegalArgumentException: No columns to insert
   at org.apache.hadoop.hbase.client.HTable.validatePut(HTable.java:950)
   at org.apache.hadoop.hbase.client.HTable.doPut(HTable.java:766)
   at org.apache.hadoop.hbase.client.HTable.put(HTable.java:752)
   at 
 org.apache.sqoop.hbase.HBasePutProcessor.accept(HBasePutProcessor.java:127)
   at 
 org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:128)
   at 
 org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:1)
   at 
 org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.write(MapTask.java:598)
   at 
 org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(TaskInputOutputContextImpl.java:89)
   at 
 org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.write(WrappedMapper.java:112)
   at 
 org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:38)
   at 
 org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:1)
   at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
   at 
 org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)
   at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:725)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
   at 
 org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:232)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
   at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
   at java.util.concurrent.FutureTask.run(FutureTask.java:138)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
   at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (SQOOP-870) Allow Sqoop to import row key column into HBase

2013-02-06 Thread David Robson (JIRA)
David Robson created SQOOP-870:
--

 Summary: Allow Sqoop to import row key column into HBase
 Key: SQOOP-870
 URL: https://issues.apache.org/jira/browse/SQOOP-870
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson


Currently when importing data into HBase, the row key column is not added to 
the column family as it is already the key. I would like to be able to tell 
Sqoop to add the row key as well.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (SQOOP-870) Allow Sqoop to import row key column into HBase

2013-02-06 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-870?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-870:
---

Attachment: SQOOP-870.patch

 Allow Sqoop to import row key column into HBase
 ---

 Key: SQOOP-870
 URL: https://issues.apache.org/jira/browse/SQOOP-870
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-870.patch


 Currently when importing data into HBase, the row key column is not added to 
 the column family as it is already the key. I would like to be able to tell 
 Sqoop to add the row key as well.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (SQOOP-862) Hbase import fails if there is a row where all columns are null

2013-02-04 Thread David Robson (JIRA)
David Robson created SQOOP-862:
--

 Summary: Hbase import fails if there is a row where all columns 
are null
 Key: SQOOP-862
 URL: https://issues.apache.org/jira/browse/SQOOP-862
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson


If you try to import a table where any of the rows contain all null values 
(except for the primary key), the import fails. For example create the 
following table in Oracle:

CREATE TABLE employee(id number primary key, test_number number);
INSERT INTO employee values(1, 123);
INSERT INTO employee values(2, null);
COMMIT;

Then run an import:

sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username USERNAME 
--table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE --column-family tst 
--hbase-create-table

The Import fails with:

 java.lang.IllegalArgumentException: No columns to insert
at 
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:400)
Caused by: java.lang.IllegalArgumentException: No columns to insert
at org.apache.hadoop.hbase.client.HTable.validatePut(HTable.java:950)
at org.apache.hadoop.hbase.client.HTable.doPut(HTable.java:766)
at org.apache.hadoop.hbase.client.HTable.put(HTable.java:752)
at 
org.apache.sqoop.hbase.HBasePutProcessor.accept(HBasePutProcessor.java:127)
at 
org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:128)
at 
org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:1)
at 
org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.write(MapTask.java:598)
at 
org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(TaskInputOutputContextImpl.java:89)
at 
org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.write(WrappedMapper.java:112)
at 
org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:38)
at 
org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:1)
at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
at 
org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)
at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:725)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
at 
org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:232)
at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (SQOOP-862) Hbase import fails if there is a row where all columns are null

2013-02-04 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-862?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-862:
---

Attachment: SQOOP-862.patch

 Hbase import fails if there is a row where all columns are null
 ---

 Key: SQOOP-862
 URL: https://issues.apache.org/jira/browse/SQOOP-862
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-862.patch


 If you try to import a table where any of the rows contain all null values 
 (except for the primary key), the import fails. For example create the 
 following table in Oracle:
 CREATE TABLE employee(id number primary key, test_number number);
 INSERT INTO employee values(1, 123);
 INSERT INTO employee values(2, null);
 COMMIT;
 Then run an import:
 sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username 
 USERNAME --table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE 
 --column-family tst --hbase-create-table
 The Import fails with:
  java.lang.IllegalArgumentException: No columns to insert
   at 
 org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:400)
 Caused by: java.lang.IllegalArgumentException: No columns to insert
   at org.apache.hadoop.hbase.client.HTable.validatePut(HTable.java:950)
   at org.apache.hadoop.hbase.client.HTable.doPut(HTable.java:766)
   at org.apache.hadoop.hbase.client.HTable.put(HTable.java:752)
   at 
 org.apache.sqoop.hbase.HBasePutProcessor.accept(HBasePutProcessor.java:127)
   at 
 org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:128)
   at 
 org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:1)
   at 
 org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.write(MapTask.java:598)
   at 
 org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(TaskInputOutputContextImpl.java:89)
   at 
 org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.write(WrappedMapper.java:112)
   at 
 org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:38)
   at 
 org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:1)
   at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
   at 
 org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)
   at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:725)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
   at 
 org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:232)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
   at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
   at java.util.concurrent.FutureTask.run(FutureTask.java:138)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
   at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (SQOOP-862) Hbase import fails if there is a row where all columns are null

2013-02-04 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-862?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13571013#comment-13571013
 ] 

David Robson commented on SQOOP-862:


Added defensive code to skip the HBase put and warn the user if it has no 
columns and would therefore fail.
Added a new unit test for this issue.

 Hbase import fails if there is a row where all columns are null
 ---

 Key: SQOOP-862
 URL: https://issues.apache.org/jira/browse/SQOOP-862
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-862.patch


 If you try to import a table where any of the rows contain all null values 
 (except for the primary key), the import fails. For example create the 
 following table in Oracle:
 CREATE TABLE employee(id number primary key, test_number number);
 INSERT INTO employee values(1, 123);
 INSERT INTO employee values(2, null);
 COMMIT;
 Then run an import:
 sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username 
 USERNAME --table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE 
 --column-family tst --hbase-create-table
 The Import fails with:
  java.lang.IllegalArgumentException: No columns to insert
   at 
 org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:400)
 Caused by: java.lang.IllegalArgumentException: No columns to insert
   at org.apache.hadoop.hbase.client.HTable.validatePut(HTable.java:950)
   at org.apache.hadoop.hbase.client.HTable.doPut(HTable.java:766)
   at org.apache.hadoop.hbase.client.HTable.put(HTable.java:752)
   at 
 org.apache.sqoop.hbase.HBasePutProcessor.accept(HBasePutProcessor.java:127)
   at 
 org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:128)
   at 
 org.apache.sqoop.mapreduce.DelegatingOutputFormat$DelegatingRecordWriter.write(DelegatingOutputFormat.java:1)
   at 
 org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.write(MapTask.java:598)
   at 
 org.apache.hadoop.mapreduce.task.TaskInputOutputContextImpl.write(TaskInputOutputContextImpl.java:89)
   at 
 org.apache.hadoop.mapreduce.lib.map.WrappedMapper$Context.write(WrappedMapper.java:112)
   at 
 org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:38)
   at 
 org.apache.sqoop.mapreduce.HBaseImportMapper.map(HBaseImportMapper.java:1)
   at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
   at 
 org.apache.sqoop.mapreduce.AutoProgressMapper.run(AutoProgressMapper.java:64)
   at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:725)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:332)
   at 
 org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:232)
   at 
 java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441)
   at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
   at java.util.concurrent.FutureTask.run(FutureTask.java:138)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
   at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
   at java.lang.Thread.run(Thread.java:662)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (SQOOP-830) HBase import formatting BigDecimal inconsistently

2013-01-30 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-830:
---

Attachment: SQOOP-830.patch

 HBase import formatting BigDecimal inconsistently
 -

 Key: SQOOP-830
 URL: https://issues.apache.org/jira/browse/SQOOP-830
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-830.patch, SQOOP-830.patch


 When importing into HBase the toString() method is called on every field via 
 the ToStringPutTransformer class.
 When the field is mapped as a BigDecimal - as it is with number fields in 
 Oracle - this results in inconsistent formats in HBase.
 For example - create the following in Oracle:
 CREATE TABLE employee(id number primary key, test_number number);
 INSERT INTO employee values(1, 0.01);
 INSERT INTO employee values(2, 0.001);
 COMMIT;
 Then run an import:
 sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username 
 USERNAME --table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE 
 --column-family tst --hbase-create-table
 The value for row 1 is 0.01 while row 2 is 1E-7.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (SQOOP-830) HBase import formatting BigDecimal inconsistently

2013-01-27 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-830:
---

Attachment: SQOOP-830.patch

 HBase import formatting BigDecimal inconsistently
 -

 Key: SQOOP-830
 URL: https://issues.apache.org/jira/browse/SQOOP-830
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Attachments: SQOOP-830.patch


 When importing into HBase the toString() method is called on every field via 
 the ToStringPutTransformer class.
 When the field is mapped as a BigDecimal - as it is with number fields in 
 Oracle - this results in inconsistent formats in HBase.
 For example - create the following in Oracle:
 CREATE TABLE employee(id number primary key, test_number number);
 INSERT INTO employee values(1, 0.01);
 INSERT INTO employee values(2, 0.001);
 COMMIT;
 Then run an import:
 sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username 
 USERNAME --table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE 
 --column-family tst --hbase-create-table
 The value for row 1 is 0.01 while row 2 is 1E-7.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (SQOOP-830) HBase import formatting BigDecimal inconsistently

2013-01-23 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson reassigned SQOOP-830:
--

Assignee: David Robson

 HBase import formatting BigDecimal inconsistently
 -

 Key: SQOOP-830
 URL: https://issues.apache.org/jira/browse/SQOOP-830
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson

 When importing into HBase the toString() method is called on every field via 
 the ToStringPutTransformer class.
 When the field is mapped as a BigDecimal - as it is with number fields in 
 Oracle - this results in inconsistent formats in HBase.
 For example - create the following in Oracle:
 CREATE TABLE employee(id number primary key, test_number number);
 INSERT INTO employee values(1, 0.01);
 INSERT INTO employee values(2, 0.001);
 COMMIT;
 Then run an import:
 sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username 
 USERNAME --table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE 
 --column-family tst --hbase-create-table
 The value for row 1 is 0.01 while row 2 is 1E-7.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (SQOOP-830) HBase import formatting BigDecimal inconsistently

2013-01-15 Thread David Robson (JIRA)
David Robson created SQOOP-830:
--

 Summary: HBase import formatting BigDecimal inconsistently
 Key: SQOOP-830
 URL: https://issues.apache.org/jira/browse/SQOOP-830
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson


When importing into HBase the toString() method is called on every field via 
the ToStringPutTransformer class.
When the field is mapped as a BigDecimal - as it is with number fields in 
Oracle - this results in inconsistent formats in HBase.
For example - create the following in Oracle:

CREATE TABLE employee(id number primary key, test_number number);
INSERT INTO employee values(1, 0.01);
INSERT INTO employee values(2, 0.001);
COMMIT;

Then run an import:

sqoop import --connect jdbc:oracle:thin:@//HOSTNAME/SERVICE --username USERNAME 
--table EMPLOYEE --password PASSWORD --hbase-table EMPLOYEE --column-family tst 
--hbase-create-table

The value for row 1 is 0.01 while row 2 is 1E-7.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (SQOOP-382) Connection parameters should be used on the mapper

2012-11-26 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson updated SQOOP-382:
---

Attachment: SQOOP-382.patch

 Connection parameters should be used on the mapper
 --

 Key: SQOOP-382
 URL: https://issues.apache.org/jira/browse/SQOOP-382
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.3

 Attachments: SQOOP-382.patch


 Currently you can specify connection parameters using --connection-param-file 
 properties-file.
 This applies the connection parameters to the connection when generating the 
 Sqoop code - but the parameters are not passed down to the mapper.
 Instead of specifying a parameters file couldn't we have a comma seperated 
 list that could be specified on the command line or in sqoop-site.xml - that 
 way it would be easier to override the settings per job, and they would be 
 passed down to the mappers. It would then be simple to modify 
 DBConfiguration.getConnection to read these.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (SQOOP-382) Connection parameters should be used on the mapper

2012-11-26 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13504273#comment-13504273
 ] 

David Robson commented on SQOOP-382:


Thanks Jarcec - for anyone using the Oracle Connector I have fixed it there as 
well: https://questmos.jira.com/browse/ORAOOP-18
Jan - The Microsoft connector may need to be updated as well - you might want 
to point them at this issue so they can fix it as well.

 Connection parameters should be used on the mapper
 --

 Key: SQOOP-382
 URL: https://issues.apache.org/jira/browse/SQOOP-382
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.3

 Attachments: SQOOP-382.patch


 Currently you can specify connection parameters using --connection-param-file 
 properties-file.
 This applies the connection parameters to the connection when generating the 
 Sqoop code - but the parameters are not passed down to the mapper.
 Instead of specifying a parameters file couldn't we have a comma seperated 
 list that could be specified on the command line or in sqoop-site.xml - that 
 way it would be easier to override the settings per job, and they would be 
 passed down to the mappers. It would then be simple to modify 
 DBConfiguration.getConnection to read these.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (SQOOP-382) Connection parameters should be used on the mapper

2012-11-25 Thread David Robson (JIRA)

 [ 
https://issues.apache.org/jira/browse/SQOOP-382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

David Robson reassigned SQOOP-382:
--

Assignee: David Robson  (was: Jarek Jarcec Cecho)

 Connection parameters should be used on the mapper
 --

 Key: SQOOP-382
 URL: https://issues.apache.org/jira/browse/SQOOP-382
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.3


 Currently you can specify connection parameters using --connection-param-file 
 properties-file.
 This applies the connection parameters to the connection when generating the 
 Sqoop code - but the parameters are not passed down to the mapper.
 Instead of specifying a parameters file couldn't we have a comma seperated 
 list that could be specified on the command line or in sqoop-site.xml - that 
 way it would be easier to override the settings per job, and they would be 
 passed down to the mappers. It would then be simple to modify 
 DBConfiguration.getConnection to read these.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (SQOOP-382) Connection parameters should be used on the mapper

2012-11-25 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13503621#comment-13503621
 ] 

David Robson commented on SQOOP-382:


Added review request https://reviews.apache.org/r/8221/

 Connection parameters should be used on the mapper
 --

 Key: SQOOP-382
 URL: https://issues.apache.org/jira/browse/SQOOP-382
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: David Robson
 Fix For: 1.4.3


 Currently you can specify connection parameters using --connection-param-file 
 properties-file.
 This applies the connection parameters to the connection when generating the 
 Sqoop code - but the parameters are not passed down to the mapper.
 Instead of specifying a parameters file couldn't we have a comma seperated 
 list that could be specified on the command line or in sqoop-site.xml - that 
 way it would be easier to override the settings per job, and they would be 
 passed down to the mappers. It would then be simple to modify 
 DBConfiguration.getConnection to read these.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (SQOOP-382) Connection parameters should be used on the mapper

2012-11-22 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13502968#comment-13502968
 ] 

David Robson commented on SQOOP-382:


It looks like you need to add the jar to the classpath - eg java TestAzure -cp 
./:./sqljdbc4.jar
You also probably need to add the connection string to your Java code to 
actually create the connection.
As for your problem - I asked a colleague who has used Sqoop on Azure before 
and he gave me the following command line he used which worked successfully:
sqoop import --connect 
jdbc:sqlserver://YOUR_HOST.database.windows.net:1433;database=YOUR_DATABASE;user=YOUR_USER@YOUR_HOST;password=YOUR_PASSWORD
 --table YOUR_TABLE --target-dir YOUR_DIRECTORY --split-by YOUR_SPLIT_COLUMN
Could you try this command - so get rid of the connection parameters completely 
- and specify the password in the URL just to see if it works?
From my experience the connection parameters using the connection-param-file 
do not work (hence this bug) - so probably best to eliminate them all together.

 Connection parameters should be used on the mapper
 --

 Key: SQOOP-382
 URL: https://issues.apache.org/jira/browse/SQOOP-382
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson

 Currently you can specify connection parameters using --connection-param-file 
 properties-file.
 This applies the connection parameters to the connection when generating the 
 Sqoop code - but the parameters are not passed down to the mapper.
 Instead of specifying a parameters file couldn't we have a comma seperated 
 list that could be specified on the command line or in sqoop-site.xml - that 
 way it would be easier to override the settings per job, and they would be 
 passed down to the mappers. It would then be simple to modify 
 DBConfiguration.getConnection to read these.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (SQOOP-382) Connection parameters should be used on the mapper

2012-11-22 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13502978#comment-13502978
 ] 

David Robson commented on SQOOP-382:


It looks to me like your problem might be with the Microsoft connector itself - 
I am not sure where to get support for that as I am not familiar with it - but 
seems like a bug they should fix.
Anyway - in regards to SQOOP-382 the connection parameters don't work - should 
this be fixed in 1.x branch or should we leave it to 2.x? If it's not going to 
be fixed in 1.x should we at least update the documentation warning people is 
does not work?

 Connection parameters should be used on the mapper
 --

 Key: SQOOP-382
 URL: https://issues.apache.org/jira/browse/SQOOP-382
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson

 Currently you can specify connection parameters using --connection-param-file 
 properties-file.
 This applies the connection parameters to the connection when generating the 
 Sqoop code - but the parameters are not passed down to the mapper.
 Instead of specifying a parameters file couldn't we have a comma seperated 
 list that could be specified on the command line or in sqoop-site.xml - that 
 way it would be easier to override the settings per job, and they would be 
 passed down to the mappers. It would then be simple to modify 
 DBConfiguration.getConnection to read these.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (SQOOP-382) Connection parameters should be used on the mapper

2012-11-22 Thread David Robson (JIRA)

[ 
https://issues.apache.org/jira/browse/SQOOP-382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13502989#comment-13502989
 ] 

David Robson commented on SQOOP-382:


I actually looked into this a fair bit when I raised it - I think the easiest 
way is to just make it a Configuration parameter - then you can use the 
standard Hadoop Configuration object to get the array of connection parameters.
Part of the problem with this for Sqoop 1.x is each connector has to implement 
it - so while it can be fixed in Sqoop - it won't work in say the Microsoft 
connector or the Quest one. Of course if it's fixed I can fix the Quest 
connector...
Let me know if this feature is desirable - seeing as I'd have to fix OraOop 
anyway I could fix it in Sqoop at the same time and submit a patch.

 Connection parameters should be used on the mapper
 --

 Key: SQOOP-382
 URL: https://issues.apache.org/jira/browse/SQOOP-382
 Project: Sqoop
  Issue Type: Bug
Reporter: David Robson
Assignee: Jarek Jarcec Cecho
 Fix For: 1.4.3


 Currently you can specify connection parameters using --connection-param-file 
 properties-file.
 This applies the connection parameters to the connection when generating the 
 Sqoop code - but the parameters are not passed down to the mapper.
 Instead of specifying a parameters file couldn't we have a comma seperated 
 list that could be specified on the command line or in sqoop-site.xml - that 
 way it would be easier to override the settings per job, and they would be 
 passed down to the mappers. It would then be simple to modify 
 DBConfiguration.getConnection to read these.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira