Hadoop 数据迁移工具,Apache Sqoop 1.4.6 发布
Sqoop是一个用来将Hadoop和关系型数据库中的数据相互转移的工具,可以将一个关系型数据库(例如 : MySQL ,Oracle ,Postgres等)中的数据导入到Hadoop的HDFS中,也可以将HDFS的数据导入到关系型数据库中。
Apache Sqoop 1.4.6 发布,这是 Apache Sqoop 地五个版本,现已提供下载:http://www.apache.org/dyn/closer.cgi/sqoop/
更新内容如下:
Bug 修复
-
[SQOOP-1125] - Out of memory errors when number of records to import < 0.5 * splitSize
-
[SQOOP-1368] - the configuration properties are reset in HBaseImportJob
-
[SQOOP-1387] - Incorrect permissions on manager.d directory can lead to NPE
-
[SQOOP-1400] - Failed to import data using mysql-connector-java-5.1.17.jar
-
[SQOOP-1411] - The number of tasks is not set properly in PGBulkloadExportManager
-
[SQOOP-1412] - Text splitter should also consider NCHAR and NVARCHAR fields
-
[SQOOP-1422] - Integration tests for Oracle connector fail as not using direct option
-
[SQOOP-1423] - hcatalog export with --map-column-java fails
-
[SQOOP-1429] - Fix native characters usage for SqlServer object names
-
[SQOOP-1433] - Oracle direct connector fails with Oracle 12c JDBC driver
-
[SQOOP-1437] - 'Text' reserved word in compliation
-
[SQOOP-1472] - Use Properties.load() method to load property files under conf/manager.d
-
[SQOOP-1474] - Fix usage of StringUtils.repeat method
-
[SQOOP-1490] - Connectors documentation doesn't build on CentOS 5
-
[SQOOP-1494] - Fix generateArvoSchema in DataDrivenImportJob
-
[SQOOP-1519] - Enable HCat/HBase/Accumulo operations with OraOop connection manager
-
[SQOOP-1520] - The table is null when using import into hive as parquet file and query option
-
[SQOOP-1524] - Error to import into hive as textfile on hive 0.13.0
-
[SQOOP-1539] - Empty partition keys and values in multiple partition key hcatalog usage should be validated
-
[SQOOP-1540] - Accumulo unit tests fail with Accumulo 1.6.1 because of conflicts in libthrift libraries
-
[SQOOP-1617] - MySQL fetch-size behavior changed with SQOOP-1400
-
[SQOOP-1627] - Fix Hadoop100 and Hadoop20 profile
-
[SQOOP-1631] - Drop confusing use of --clean-staging-table parameter from PGBulkloadManager
-
[SQOOP-1663] - OraOop test cases are not logging any output
-
[SQOOP-1682] - Test cases *LobAvroImportTest are failing
-
[SQOOP-1684] - Use pre-existing HBase delegation token
-
[SQOOP-1685] - HCatalog integration is not working on JDK8
-
[SQOOP-1759] - TestIncrementalImport fails with NPE on Windows
-
[SQOOP-1764] - Numeric Overflow when getting extent map
-
[SQOOP-1779] - Add support for --hive-database when importing Parquet files into Hive
-
[SQOOP-1826] - NPE in ImportTool.lastModifiedMerge during postgres import
-
[SQOOP-1890] - Properly escape table name in generated queries
-
[SQOOP-1970] - Add warning about trailing whitespace characters when using password file to User guide
-
[SQOOP-2017] - Print out loaded columns and their type in verbose mode
-
[SQOOP-2024] - Hive import doesn't remove target directory in hive
-
[SQOOP-2055] - Run only one map task attempt during export
-
[SQOOP-2057] - Skip delegation token generation flag during hbase import
-
[SQOOP-2128] - HBaseImportJob should close connection from HBaseAdmin to HBase
-
[SQOOP-2130] - BaseSqoopTestCase should use manager.escapeTable instead of directly hardcoding double quotes
-
[SQOOP-2132] - Remove test TestMainframeImportTool.testPrintHelp
-
[SQOOP-2136] - Test case SqlServerUpsertOutputFormatTest is failing
-
[SQOOP-2137] - Sqoop tests and documentation refer to as-avrofile (instead of as-avrodatafile)
-
[SQOOP-2145] - Default Hive home is not being set properly under certain circumstances
-
[SQOOP-2164] - Enhance the Netezza Connector for Sqoop
-
[SQOOP-2170] - MySQL specific tests are not properly cleaning up created tables
-
[SQOOP-2191] - Provide an option automatically choose one mapper when neither primary key is defined nor split by column is provided
-
[SQOOP-2254] - Failed to build release notes
-
[SQOOP-2257] - Parquet target for imports with Hive overwrite option does not work
-
[SQOOP-2263] - Sqoop1 has some files without a copyright header
-
[SQOOP-2264] - Exclude and remove SqoopUserGuide.xml from git repository
-
[SQOOP-2281] - Set overwrite on kite dataset
-
[SQOOP-2282] - Add validation check for --hive-import and --append
-
[SQOOP-2283] - Support usage of --exec and --password-alias
-
[SQOOP-2286] - Ensure Sqoop generates valid avro column names
-
[SQOOP-2290] - java.lang.ArrayIndexOutOfBoundsException thrown when malformed column mapping is provided
-
[SQOOP-2294] - Change to Avro schema name breaks some use cases
-
[SQOOP-2324] - Remove extra license handling for consistency
改进
-
[SQOOP-1330] - Ignore blank newlines in managers.d property files
-
[SQOOP-1391] - Compression codec handling
-
[SQOOP-1392] - Create the temporary directory inside task working dir rather then in tmp
-
[SQOOP-1421] - Automated patch script
-
[SQOOP-1471] - Use Hadoop CredentialProvider API to encyrpt passwords at rest
-
[SQOOP-1489] - Propagate cubrid properties to the test VM
-
[SQOOP-1567] - Auto-Configure JTDS Driver From JDBCUrl
-
[SQOOP-1622] - Copying from staging table should be in single transaction for pg_bulkload connector
-
[SQOOP-1632] - Add support for index organized tables to direct connector
-
[SQOOP-2149] - Update Kite dependency to 1.0.0
-
[SQOOP-2252] - Add default to Avro Schema
新特性
-
[SQOOP-1272] - Support importing mainframe sequential datasets
-
[SQOOP-1309] - Expand the Sqoop to support CUBRID database.
-
[SQOOP-1366] - Propose to add Parquet support
-
[SQOOP-1403] - Upsert export for SQL Server
-
[SQOOP-1405] - Add arg to enable SQL Server identity insert on export
-
[SQOOP-1450] - Copy the avro schema file to hdfs for AVRO based import
更多更新内容请看:
https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12327469&projectId=12311320