Repository: sqoop Updated Branches: refs/heads/trunk b0f391e75 -> 8929e61fb
SQOOP-3202: Update change log with 1.4.7 release (Attila Szabo) Project: http://git-wip-us.apache.org/repos/asf/sqoop/repo Commit: http://git-wip-us.apache.org/repos/asf/sqoop/commit/8929e61f Tree: http://git-wip-us.apache.org/repos/asf/sqoop/tree/8929e61f Diff: http://git-wip-us.apache.org/repos/asf/sqoop/diff/8929e61f Branch: refs/heads/trunk Commit: 8929e61fb4d0198d79c612c1999f1071e505434f Parents: b0f391e Author: Attila Szabo <[email protected]> Authored: Fri Jun 16 07:52:47 2017 +0200 Committer: Attila Szabo <[email protected]> Committed: Fri Jun 16 07:53:41 2017 +0200 ---------------------------------------------------------------------- CHANGELOG.txt | 206 +++++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 206 insertions(+) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/sqoop/blob/8929e61f/CHANGELOG.txt ---------------------------------------------------------------------- diff --git a/CHANGELOG.txt b/CHANGELOG.txt index 5bbac09..a67f430 100644 --- a/CHANGELOG.txt +++ b/CHANGELOG.txt @@ -1,3 +1,209 @@ +Release Notes - Sqoop - Version 1.4.7 + +** Sub-task + * [SQOOP-956] - Create a winpkg target for Sqoop to create a Winows installable Sqoop package + * [SQOOP-961] - Update Sqoop documentation for Windows changes + * [SQOOP-2643] - Incremental imports fail in Sqoop when run using Teradata JDBC driver + * [SQOOP-2644] - Sqoop Export does not offer upserts from HDFS to Teradata DB + * [SQOOP-2937] - Sqoop mainframe module does not support sequential data sets, GDG + * [SQOOP-2938] - Mainframe import module extension to support data sets on tape + * [SQOOP-3055] - SQOOP-3055 MYSQL tests are failing due to the tests ignoring specified username, password and dbname, trying to connect to specified host using "currentUser" + * [SQOOP-3056] - Add ant ivy report target + * [SQOOP-3057] - Fix failing 3rd party Oracle tests + * [SQOOP-3091] - Clean up expected exception logic in tests - part I. + * [SQOOP-3092] - Clean up expected exception logic in tests - part II. + * [SQOOP-3141] - Fix Java7 compile issue (getTypeName is only available after Java 8) + * [SQOOP-3142] - Restore fail messages removed in SQOOP-3092 + * [SQOOP-3143] - Restore fail messages removed in SQOOP-3091 + +** Bug + * [SQOOP-890] - ClassWriter Generates setField Casts Based on the Output Type, not the Input Type + * [SQOOP-1198] - Make Sqoop build aware of the protobuf library + * [SQOOP-1252] - Sqoop import from db2. Schema of import not in expected format. + * [SQOOP-1289] - Sqoop hbase-bulkload does not work with composite key + * [SQOOP-1295] - Sqoop Merge supports only one column merge key. Merge fails if source table as composite key + * [SQOOP-1301] - 'run' method in Sqoop.java is not thread-safe + * [SQOOP-1347] - High performance oracle connector should depend on --direct flag only to disable/enable feature + * [SQOOP-1369] - Avro export ignores --columns option + * [SQOOP-1493] - Add ability to import/export true decimal in Avro instead of serializing it to String + * [SQOOP-1600] - Exception when import data using Data Connector for Oracle with TIMESTAMP column type to Parquet files + * [SQOOP-1735] - Sqoop job fails (but not 'sqoop import') if --create-hive-table is set and the hive table already exists + * [SQOOP-1760] - Relocated hadoop installation + * [SQOOP-1807] - Incremental import to HBase with free form query is broken + * [SQOOP-1932] - fix authorization failuare when the --hive-table option contaion a database name for import + * [SQOOP-1933] - CryptoFileLoader does not work for saved jobs + * [SQOOP-2103] - Not able define Decimal(n,p) data type in map-column-hive option + * [SQOOP-2264] - Exclude and remove SqoopUserGuide.xml from git repository + * [SQOOP-2283] - Support usage of --exec and --password-alias + * [SQOOP-2286] - Ensure Sqoop generates valid avro column names + * [SQOOP-2295] - Hive import with Parquet should append automatically + * [SQOOP-2296] - Merge statement incorrectly includes table hints + * [SQOOP-2297] - Explicitly add zookeeper as a dependency in ivy.xml + * [SQOOP-2298] - TestAvroImport test case error + * [SQOOP-2326] - Fix Netezza trunc-string option handling and unnecessary log directory during imports + * [SQOOP-2328] - Sqoop import does not recognize Primary Key of a IBM DB2 table + * [SQOOP-2339] - Move sub-directory might fail in append mode + * [SQOOP-2343] - AsyncSqlRecordWriter stucks if any exception is thrown out in its close method + * [SQOOP-2346] - compression isn't honored in incremental imports + * [SQOOP-2349] - Transaction isolation level for metadata queries should be mutable + * [SQOOP-2362] - Add oracle direct mode in list of supported databases + * [SQOOP-2363] - wrong option for escape character with mysqlimport + * [SQOOP-2367] - Tests from PostgresqlImportTest can fail on various versions of Hadoop + * [SQOOP-2370] - Netezza - need to support additional options for full control character handling + * [SQOOP-2371] - Tests from *LobAvroImportTest might fail + * [SQOOP-2372] - Imports all tables as parquet will meet a NPE + * [SQOOP-2381] - Add test for mysql export with --escape-by option + * [SQOOP-2387] - Sqoop should support importing from table with column names containing some special character + * [SQOOP-2399] - BigDecimalSplitter java.lang.ArrayIndexOutOfBoundsException + * [SQOOP-2400] - hive.metastore.sasl.enabled should be set to true for Oozie integration + * [SQOOP-2406] - Add support for secure mode when importing Parquet files into Hive + * [SQOOP-2437] - Use hive configuration to connect to secure metastore + * [SQOOP-2438] - Use Class.cast when creating HiveConf object in ParquetJob + * [SQOOP-2454] - Drop JDK6 support + * [SQOOP-2470] - Incremental Hive import with append not working after validation check for --hive-import and --import + * [SQOOP-2531] - readlink -f not supported on OS X + * [SQOOP-2535] - Add error handling to HiveConf + * [SQOOP-2561] - Special Character removal from Column name as avro data results in duplicate column and fails the import + * [SQOOP-2566] - Sqoop incremental import using --query switch throws NullPointerException with a managed table + * [SQOOP-2582] - Query import won't work for parquet + * [SQOOP-2596] - Precision of varchar/char column cannot be retrieved from teradata database during sqoop import + * [SQOOP-2597] - Missing method AvroSchemaGenerator.generate() + * [SQOOP-2607] - Direct import from Netezza and encoding + * [SQOOP-2617] - Log file that is being used for LOB files + * [SQOOP-2620] - Sqoop fails with Incremental Imports and Upserts in Teradata using Teradata JDBC Driver + * [SQOOP-2627] - Incremental imports fail in Sqoop when run using Teradata JDBC driver + * [SQOOP-2642] - Document ability to specify commas in --map-column-hive option + * [SQOOP-2651] - Do not dump data on error in TextExportMapper by default + * [SQOOP-2707] - Upgrade commons-collections to 3.2.2 + * [SQOOP-2712] - Run only one map task attempt during export (second edition) + * [SQOOP-2723] - Oracle connector not working with lowercase columns + * [SQOOP-2737] - Cannot import table from Oracle with column with spaces in name + * [SQOOP-2745] - Using datetime column as a splitter for Oracle no longer works + * [SQOOP-2746] - Add test case for Oracle incremental import using Timestamp + * [SQOOP-2747] - Allow customizing test username and password for Oracle tests + * [SQOOP-2751] - Test TestIncrementalImport is failing + * [SQOOP-2753] - TestSqoopJsonUtil.testGetJsonStringFromMap is depending on Map ordering in JDK + * [SQOOP-2767] - Test is failing SystemImportTest + * [SQOOP-2779] - Sqoop metastore doesn't seem to recognize --schema option + * [SQOOP-2780] - Sqoop 1 unit tests fail with TestIncrementalImport test + * [SQOOP-2783] - Query import with parquet fails on incompatible schema + * [SQOOP-2787] - MySql import and export fails with 5.1 server and 5.1.17+ drivers + * [SQOOP-2810] - Upgrade to non-snapshot dependency on Avro 1.8.0 as soon as it gets released + * [SQOOP-2815] - Add documentation for Atlas Integration + * [SQOOP-2839] - Sqoop import failure due to data member conflict in ORM code for table + * [SQOOP-2846] - Sqoop Export with update-key failing for avro data file + * [SQOOP-2847] - Sqoop --incremental + missing parent --target-dir reports success with no data + * [SQOOP-2850] - Append mode for hive imports is not yet supported. Please remove the parameter --append-mode + * [SQOOP-2858] - Sqoop export with Avro data using (--update-key <key> and --update-mode allowinsert) fails + * [SQOOP-2863] - Properly escape column names for generated INSERT statements + * [SQOOP-2864] - ClassWriter chokes on column names containing double quotes + * [SQOOP-2880] - Provide argument for overriding temporary directory + * [SQOOP-2884] - Document --temporary-rootdir + * [SQOOP-2894] - Hive import with Parquet failed in Kerberos enabled cluster + * [SQOOP-2896] - Sqoop exec job fails with SQLException Access denied for user + * [SQOOP-2909] - Oracle related ImportTest fails after SQOOP-2737 + * [SQOOP-2911] - Fix failing HCatalogExportTest caused by SQOOP-2863 + * [SQOOP-2915] - Fixing Oracle related unit tests + * [SQOOP-2920] - sqoop performance deteriorates significantly on wide datasets; sqoop 100% on cpu + * [SQOOP-2936] - Provide Apache Atlas integration for hcatalog based exports + * [SQOOP-2945] - Oracle CLOB mapped to String is unable to import the data (SQLException in nextKeyValue) + * [SQOOP-2946] - Netezza export fails when BOOLEAN column is mapped with INTEGER + * [SQOOP-2947] - Oracle direct mode do not allow --jar-file to have fewer columns to export the data + * [SQOOP-2950] - Sqoop trunk has consistent UT failures - need fixing + * [SQOOP-2952] - row key not added into column family using --hbase-bulkload + * [SQOOP-2971] - OraOop does not close connections properly + * [SQOOP-2978] - Netezza import/export fails when TIME column is mapped with TIMESTAMP + * [SQOOP-2979] - Oracle direct mode do not allow FLOAT data type (java.lang.ClassCastException: java.lang.Double cannot be cast to java.math.BigDecimal) + * [SQOOP-2980] - Export to DB2 z/OS fails unless --batch mode is used + * [SQOOP-2983] - OraOop export has degraded performance with wide tables + * [SQOOP-2986] - Add validation check for --hive-import and --incremental lastmodified + * [SQOOP-2990] - Sqoop(oracle) export [updateTableToOracle] with "--update-mode allowinsert" : app fails with java.sql.SQLException: Missing IN or OUT parameter at index + * [SQOOP-2995] - backward incompatibility introduced by Custom Tool options + * [SQOOP-2999] - Sqoop ClassNotFoundException (org.apache.commons.lang3.StringUtils) is thrown when executing Oracle direct import map task + * [SQOOP-3010] - Sqoop should not allow --as-parquetfile with hcatalog jobs or when hive import with create-hive-table is used + * [SQOOP-3013] - Configuration "tmpjars" is not checked for empty strings before passing to MR + * [SQOOP-3014] - Sqoop with HCatalog import loose precision for large numbers that does not fit into double + * [SQOOP-3021] - ClassWriter fails if a column name contains a backslash character + * [SQOOP-3033] - Sqoop option --skip-dist-cache is not saved as a parameter when saving Sqoop Job + * [SQOOP-3038] - Sqoop export using --hcatalog with RDBMS reserved word column name results in "null" value + * [SQOOP-3044] - Add missing ASF license information to .java files + * [SQOOP-3054] - Get FileSystem from parameter "--target-dir" + * [SQOOP-3061] - Sqoop --options-file failed with error "Malformed option in options file" even though the query is correct + * [SQOOP-3069] - Get OracleExportTest#testUpsertTestExport in line with SQOOP-3066 + * [SQOOP-3071] - Fix OracleManager to apply localTimeZone correctly in case of Date objects too + * [SQOOP-3072] - Reenable escaping in ImportTest#testProductWithWhiteSpaceImport for proper execution + * [SQOOP-3074] - Fix Avro import not to fail with Javac errors in case of non UTF-8 locale + * [SQOOP-3081] - use OracleEscapeUtils.escapeIdentifier OracleUpsertOutputFormat + add compatibility with SQOOP-3066 + * [SQOOP-3083] - Fault injection targets no longer work after SQOOP-2983 + * [SQOOP-3105] - Postgres Tests should clean up after themselves + * [SQOOP-3123] - Import from oracle using oraoop with map-column-java to avro fails if special characters encounter in table name or column name + * [SQOOP-3124] - Fix ordering in column list query of PostgreSQL connector + * [SQOOP-3127] - Increase timeout in TestClassWriter#testWideTableClassGeneration to avoid flaky test scenarios in the upstream Jenkins + * [SQOOP-3138] - Netezza Direct Import does not support --columns options + * [SQOOP-3140] - mapred.map.max.attempts is deprecated. Instead, use mapreduce.map.maxattempts - old property is used by SQOOP-2055 + * [SQOOP-3152] - --map-column-hive to support DECIMAL(xx,xx) + * [SQOOP-3157] - Improve regex introduced in [SQOOP-3152] + * [SQOOP-3159] - Sqoop (export + --table) with Oracle table_name having '$' fails with error (ORA-00942 or java.lang.NoClassDefFoundError) + * [SQOOP-3173] - support DB2 xml data type when sqoop import with parquet + +** Improvement + * [SQOOP-816] - Scoop and support for external Hive tables + * [SQOOP-957] - Proposed enhancements for getting Sqoop support on Windows + * [SQOOP-1281] - Support of glob paths during export + * [SQOOP-1904] - support for DB2 XML data type when importing to hdfs + * [SQOOP-1905] - add --schema option for import-all-tables and list-tables against db2 + * [SQOOP-1906] - Export support for mixed update/insert against db2 + * [SQOOP-1907] - export support for --staging-table against db2 + * [SQOOP-2457] - Add option to automatically compute statistics after loading date into a hive table + * [SQOOP-2647] - Add option for drop-if-exists when using sqoop hcat import. + * [SQOOP-2795] - Clean up useless code in class TestSqoopJsonUtil + * [SQOOP-2801] - Secure RDBMS password in Sqoop Metastore in a encrypted form + * [SQOOP-2906] - Optimization of AvroUtil.toAvroIdentifier + * [SQOOP-2910] - Add capability to Sqoop to require an explicit option to be specified with --split-by for a String column + * [SQOOP-2913] - Make sqoop fails if user uses --direct connector for case when --direct connector is not available + * [SQOOP-2939] - Extend mainframe module to support GDG, sequential data sets, and data sets stored on tape + * [SQOOP-2943] - Make sqoop able to import to Parquet file format in case of HDFS encryption zones are turned on + * [SQOOP-2944] - AURORA direct mode to support Sequence file format + * [SQOOP-2972] - SQOOP Direct Export To PostgreSQL Supports Selective Columns + * [SQOOP-3009] - Import comments for columns for Postgresql + * [SQOOP-3026] - Document that Sqoop export with --hcatalog-table <HIVE_VIEW> is not supported + * [SQOOP-3027] - Create check/fail fast for Sqoop export with --hcatalog-table <HIVE_VIEW>, as it's not supported by Hive + MR + * [SQOOP-3028] - Include stack trace in the logging of exceptions in ExportTool + * [SQOOP-3029] - Add an option for uppercase/lowercase column name mapping between HCatalog and RDBMS cloumn name list + * [SQOOP-3034] - HBase import should fail fast if using anything other than as-textfile + * [SQOOP-3037] - Minor convenience feature - add flag to ant test to enable remote debugging + * [SQOOP-3050] - Create an compile/execution profile which is capable of running all the available test (including the 3rd party tests) + * [SQOOP-3051] - Remove/delete obsolete profiles from build.xml + * [SQOOP-3052] - Introduce Maven/Gradle/etc. based build for Sqoop to make it more developer friendly / open + * [SQOOP-3053] - Create a cmd line argument for sqoop.throwOnError and use it through SqoopOptions + * [SQOOP-3066] - Introduce an option + env variable to enable/disable SQOOP-2737 feature + * [SQOOP-3067] - Add an cmd line option to support split-by feature for database functions/expressions + * [SQOOP-3068] - Enhance error (tool.ImportTool: Encountered IOException running import job: java.io.IOException: Expected schema) to suggest workaround (--map-column-java) + * [SQOOP-3085] - Add support for client side (JVM) timezone settings + * [SQOOP-3090] - Normalize test cases where expect an exception + * [SQOOP-3131] - Docuemtn support for DB2 XML data type when importing to hdfs + * [SQOOP-3135] - Not enough error message for debugging when parameters missing + * [SQOOP-3136] - Sqoop should work well with not default file systems + * [SQOOP-3158] - Columns added to Mysql after initial sqoop import, export back to table with same schema fails + * [SQOOP-3169] - Evaluate and fix SQLServer Manual tests + * [SQOOP-3190] - Remove dependency on PSQL for postgres direct import + * [SQOOP-3192] - upgrade parquet + +** New Feature + * [SQOOP-1094] - Add Avro support to merge tool + * [SQOOP-2331] - Snappy Compression Support in Sqoop-HCatalog + * [SQOOP-2332] - Dynamic Partition in Sqoop HCatalog- if Hive table does not exists & add support for Partition Date Format + * [SQOOP-2333] - Sqoop to support Custom options for User Defined Plugins(Tool) + * [SQOOP-2334] - Sqoop Volume Per Mapper + * [SQOOP-2335] - Support for Hive External Table in Sqoop - HCatalog + * [SQOOP-2534] - --password-file option doesn't work Teradata jdbc driver + * [SQOOP-2585] - merging hive tables using sqoop + * [SQOOP-2609] - Provide Apache Atlas integration for hive and hcatalog based imports + * [SQOOP-2649] - Support for importing data onto Apache Phoenix tables + +** Task + * [SQOOP-960] - Update Sqoop documentation for Windows changes + * [SQOOP-3080] - Correct default transaction isolation level comment in SqoopOptions Release Notes - Sqoop - Version 1.4.6
