Hadoop 数据迁移工具,Apache Sqoop 1.4.6 发布

pjp 9年前

Hadoop 数据迁移工具,Apache Sqoop 1.4.6 发布

Sqoop是一个用来将Hadoop和关系型数据库中的数据相互转移的工具,可以将一个关系型数据库(例如 : MySQL ,Oracle ,Postgres等)中的数据导入到Hadoop的HDFS中,也可以将HDFS的数据导入到关系型数据库中。

Apache Sqoop 1.4.6 发布,这是 Apache Sqoop 地五个版本,现已提供下载:

http://www.apache.org/dyn/closer.cgi/sqoop/ 

更新内容如下:

Bug 修复

  • [SQOOP-1125] - Out of memory errors when number of records to import < 0.5 * splitSize

  • [SQOOP-1368] - the configuration properties are reset in HBaseImportJob

  • [SQOOP-1387] - Incorrect permissions on manager.d directory can lead to NPE

  • [SQOOP-1400] - Failed to import data using mysql-connector-java-5.1.17.jar

  • [SQOOP-1411] - The number of tasks is not set properly in PGBulkloadExportManager

  • [SQOOP-1412] - Text splitter should also consider NCHAR and NVARCHAR fields

  • [SQOOP-1422] - Integration tests for Oracle connector fail as not using direct option

  • [SQOOP-1423] - hcatalog export with --map-column-java fails

  • [SQOOP-1429] - Fix native characters usage for SqlServer object names

  • [SQOOP-1433] - Oracle direct connector fails with Oracle 12c JDBC driver

  • [SQOOP-1437] - 'Text' reserved word in compliation

  • [SQOOP-1472] - Use Properties.load() method to load property files under conf/manager.d

  • [SQOOP-1474] - Fix usage of StringUtils.repeat method

  • [SQOOP-1490] - Connectors documentation doesn't build on CentOS 5

  • [SQOOP-1494] - Fix generateArvoSchema in DataDrivenImportJob

  • [SQOOP-1519] - Enable HCat/HBase/Accumulo operations with OraOop connection manager

  • [SQOOP-1520] - The table is null when using import into hive as parquet file and query option

  • [SQOOP-1524] - Error to import into hive as textfile on hive 0.13.0

  • [SQOOP-1539] - Empty partition keys and values in multiple partition key hcatalog usage should be validated

  • [SQOOP-1540] - Accumulo unit tests fail with Accumulo 1.6.1 because of conflicts in libthrift libraries

  • [SQOOP-1617] - MySQL fetch-size behavior changed with SQOOP-1400

  • [SQOOP-1627] - Fix Hadoop100  and Hadoop20 profile

  • [SQOOP-1631] - Drop confusing use of --clean-staging-table parameter from PGBulkloadManager

  • [SQOOP-1663] - OraOop test cases are not logging any output

  • [SQOOP-1682] - Test cases *LobAvroImportTest are failing

  • [SQOOP-1684] - Use pre-existing HBase delegation token

  • [SQOOP-1685] - HCatalog integration is not working on JDK8

  • [SQOOP-1759] - TestIncrementalImport fails with NPE on Windows

  • [SQOOP-1764] - Numeric Overflow when getting extent map

  • [SQOOP-1779] - Add support for --hive-database when importing Parquet files into Hive

  • [SQOOP-1826] - NPE in ImportTool.lastModifiedMerge during postgres import

  • [SQOOP-1890] - Properly escape table name in generated queries

  • [SQOOP-1970] - Add warning about trailing whitespace characters when using password file to User guide

  • [SQOOP-2017] - Print out loaded columns and their type in verbose mode

  • [SQOOP-2024] - Hive import doesn't remove target directory in hive

  • [SQOOP-2055] - Run only one map task attempt during export

  • [SQOOP-2057] - Skip delegation token generation flag during hbase import

  • [SQOOP-2128] - HBaseImportJob should close connection from HBaseAdmin to HBase

  • [SQOOP-2130] - BaseSqoopTestCase should use manager.escapeTable instead of directly hardcoding double quotes

  • [SQOOP-2132] - Remove test TestMainframeImportTool.testPrintHelp

  • [SQOOP-2136] - Test case SqlServerUpsertOutputFormatTest is failing

  • [SQOOP-2137] - Sqoop tests and documentation refer to as-avrofile (instead of as-avrodatafile)

  • [SQOOP-2145] - Default Hive home is not being set properly under certain circumstances

  • [SQOOP-2164] - Enhance the Netezza Connector for Sqoop

  • [SQOOP-2170] - MySQL specific tests are not properly cleaning up created tables

  • [SQOOP-2191] - Provide an option automatically choose one mapper when neither primary key is defined nor split by column is provided

  • [SQOOP-2254] - Failed to build release notes

  • [SQOOP-2257] - Parquet target for imports with Hive overwrite option does not work

  • [SQOOP-2263] - Sqoop1 has some files without a copyright header

  • [SQOOP-2264] - Exclude and remove SqoopUserGuide.xml from git repository

  • [SQOOP-2281] - Set overwrite on kite dataset

  • [SQOOP-2282] - Add validation check for --hive-import and --append

  • [SQOOP-2283] - Support usage of --exec and --password-alias

  • [SQOOP-2286] - Ensure Sqoop generates valid avro column names

  • [SQOOP-2290] - java.lang.ArrayIndexOutOfBoundsException thrown when malformed column mapping is provided

  • [SQOOP-2294] - Change to Avro schema name breaks some use cases

  • [SQOOP-2324] - Remove extra license handling for consistency

改进

  • [SQOOP-1330] - Ignore blank newlines in managers.d property files

  • [SQOOP-1391] - Compression codec handling

  • [SQOOP-1392] - Create the temporary directory inside task working dir rather then in tmp

  • [SQOOP-1421] - Automated patch script

  • [SQOOP-1471] - Use Hadoop CredentialProvider API to encyrpt passwords at rest

  • [SQOOP-1489] - Propagate cubrid properties to the test VM

  • [SQOOP-1567] - Auto-Configure JTDS Driver From JDBCUrl

  • [SQOOP-1622] - Copying from staging table should be in single transaction for pg_bulkload connector

  • [SQOOP-1632] - Add support for index organized tables to direct connector

  • [SQOOP-2149] - Update Kite dependency to 1.0.0

  • [SQOOP-2252] - Add default to Avro Schema

新特性

  • [SQOOP-1272] - Support importing mainframe sequential datasets

  • [SQOOP-1309] - Expand the Sqoop to support CUBRID database.

  • [SQOOP-1366] - Propose to add Parquet support

  • [SQOOP-1403] - Upsert export for SQL Server

  • [SQOOP-1405] - Add arg to enable SQL Server identity insert on export

  • [SQOOP-1450] - Copy the avro schema file to hdfs for AVRO based import

更多更新内容请看:

https://issues.apache.org/jira/secure/ReleaseNote.jspa?version=12327469&projectId=12311320