sqoop-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From venkatran...@apache.org
Subject [1/3] sqoop git commit: SQOOP-2745: Using datetime column as a splitter for Oracle no longer works (Jarek Jarcec Cecho via Venkat Ranganathan)
Date Thu, 17 Dec 2015 04:50:13 GMT
Repository: sqoop
Updated Branches:
  refs/heads/trunk e077f9978 -> f19e2a523


SQOOP-2745: Using datetime column as a splitter for Oracle no longer works
  (Jarek Jarcec Cecho via Venkat Ranganathan)


Project: http://git-wip-us.apache.org/repos/asf/sqoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/sqoop/commit/9c7638d7
Tree: http://git-wip-us.apache.org/repos/asf/sqoop/tree/9c7638d7
Diff: http://git-wip-us.apache.org/repos/asf/sqoop/diff/9c7638d7

Branch: refs/heads/trunk
Commit: 9c7638d74180cc607ef509b9dd9e6a45ff60c041
Parents: e077f99
Author: Venkat Ranganathan <venkat@hortonworks.com>
Authored: Wed Dec 16 20:47:59 2015 -0800
Committer: Venkat Ranganathan <venkat@hortonworks.com>
Committed: Wed Dec 16 20:47:59 2015 -0800

----------------------------------------------------------------------
 .../db/OracleDataDrivenDBInputFormat.java       |   4 +-
 .../com/cloudera/sqoop/ThirdPartyTests.java     |   2 +
 .../manager/oracle/OracleSplitterTest.java      | 147 +++++++++++++++++++
 3 files changed, 151 insertions(+), 2 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/sqoop/blob/9c7638d7/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java
----------------------------------------------------------------------
diff --git a/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java b/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java
index 15c5898..8b5103b 100644
--- a/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java
+++ b/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java
@@ -45,7 +45,7 @@ public class OracleDataDrivenDBInputFormat<T extends DBWritable>
    * into InputSplits.
    */
   @Override
-  protected DBSplitter getSplitter(int sqlDataType) {
+  protected DBSplitter getSplitter(int sqlDataType, long splitLimit) {
     switch (sqlDataType) {
     case Types.DATE:
     case Types.TIME:
@@ -53,7 +53,7 @@ public class OracleDataDrivenDBInputFormat<T extends DBWritable>
       return new OracleDateSplitter();
 
     default:
-      return super.getSplitter(sqlDataType);
+      return super.getSplitter(sqlDataType, splitLimit);
     }
   }
 

http://git-wip-us.apache.org/repos/asf/sqoop/blob/9c7638d7/src/test/com/cloudera/sqoop/ThirdPartyTests.java
----------------------------------------------------------------------
diff --git a/src/test/com/cloudera/sqoop/ThirdPartyTests.java b/src/test/com/cloudera/sqoop/ThirdPartyTests.java
index 06b4ae6..50f3192 100644
--- a/src/test/com/cloudera/sqoop/ThirdPartyTests.java
+++ b/src/test/com/cloudera/sqoop/ThirdPartyTests.java
@@ -57,6 +57,7 @@ import org.apache.sqoop.manager.netezza.NetezzaExportManualTest;
 import org.apache.sqoop.manager.netezza.NetezzaImportManualTest;
 import org.apache.sqoop.manager.oracle.OracleCallExportTest;
 import org.apache.sqoop.manager.oracle.OracleIncrementalImportTest;
+import org.apache.sqoop.manager.oracle.OracleSplitterTest;
 import org.apache.sqoop.manager.sqlserver.SQLServerDatatypeExportDelimitedFileManualTest;
 import org.apache.sqoop.manager.sqlserver.SQLServerDatatypeExportSequenceFileManualTest;
 import org.apache.sqoop.manager.sqlserver.SQLServerDatatypeImportDelimitedFileManualTest;
@@ -96,6 +97,7 @@ public final class ThirdPartyTests extends TestCase {
     suite.addTestSuite(OracleManagerTest.class);
     suite.addTestSuite(OracleCompatTest.class);
     suite.addTestSuite(OracleIncrementalImportTest.class);
+    suite.addTestSuite(OracleSplitterTest.class);
 
     // SQL Server
     suite.addTestSuite(SQLServerDatatypeExportDelimitedFileManualTest.class);

http://git-wip-us.apache.org/repos/asf/sqoop/blob/9c7638d7/src/test/org/apache/sqoop/manager/oracle/OracleSplitterTest.java
----------------------------------------------------------------------
diff --git a/src/test/org/apache/sqoop/manager/oracle/OracleSplitterTest.java b/src/test/org/apache/sqoop/manager/oracle/OracleSplitterTest.java
new file mode 100644
index 0000000..fc5f56b
--- /dev/null
+++ b/src/test/org/apache/sqoop/manager/oracle/OracleSplitterTest.java
@@ -0,0 +1,147 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.sqoop.manager.oracle;
+
+import com.cloudera.sqoop.SqoopOptions;
+import com.cloudera.sqoop.manager.OracleUtils;
+import com.cloudera.sqoop.testutil.CommonArgs;
+import com.cloudera.sqoop.testutil.ImportJobTestCase;
+import org.apache.commons.logging.Log;
+import org.apache.commons.logging.LogFactory;
+import org.apache.commons.io.FileUtils;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.FSDataInputStream;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.io.IOUtils;
+
+import java.io.BufferedReader;
+import java.io.BufferedWriter;
+import java.io.File;
+import java.io.FileInputStream;
+import java.io.FileWriter;
+import java.io.IOException;
+import java.io.InputStreamReader;
+import java.io.Writer;
+import java.sql.SQLException;
+import java.util.ArrayList;
+import java.util.List;
+
+/**
+ * Test various custom splitters for Oracle.
+ */
+public class OracleSplitterTest extends ImportJobTestCase {
+
+  public static final Log LOG = LogFactory.getLog(
+      OracleSplitterTest.class.getName());
+
+  @Override
+  protected boolean useHsqldbTestServer() {
+    return false;
+  }
+
+  @Override
+  protected String getConnectString() {
+    return OracleUtils.CONNECT_STRING;
+  }
+
+  @Override
+  protected SqoopOptions getSqoopOptions(Configuration conf) {
+    SqoopOptions opts = new SqoopOptions(conf);
+    OracleUtils.setOracleAuth(opts);
+    return opts;
+  }
+
+  @Override
+  protected void dropTableIfExists(String table) throws SQLException {
+    OracleUtils.dropTable(table, getManager());
+  }
+
+  /** the names of the tables we're creating. */
+  private List<String> tableNames;
+
+  @Override
+  public void tearDown() {
+    // Clean up the database on our way out.
+    for (String tableName : tableNames) {
+      try {
+        dropTableIfExists(tableName);
+      } catch (SQLException e) {
+        LOG.warn("Error trying to drop table '" + tableName
+                 + "' on tearDown: " + e);
+      }
+    }
+    super.tearDown();
+  }
+
+  protected String [] getArgv(String tableName, String connPropsFileName, String splitByColumn)
{
+    ArrayList<String> args = new ArrayList<String>();
+
+    CommonArgs.addHadoopFlags(args);
+
+    args.add("--connect");
+    args.add(getConnectString());
+    args.add("--target-dir");
+    args.add(getWarehouseDir());
+    args.add("--num-mappers");
+    args.add("2");
+    args.add("--split-by");
+    args.add(splitByColumn);
+    args.add("--table");
+    args.add(tableName);
+    args.add("--connection-param-file");
+    args.add(connPropsFileName);
+
+    return args.toArray(new String[0]);
+  }
+
+  public void testTimestampSplitter() throws IOException {
+    tableNames = new ArrayList<String>();
+    String [] types = { "INT", "VARCHAR(10)", "TIMESTAMP", };
+    String [] vals = {
+      "1", "'old_data'", "TO_TIMESTAMP('1999-01-01 11:11:11', 'YYYY-MM-DD HH24:MI:SS')",
+      "2", "'new_data'", "TO_TIMESTAMP('2000-11-11 23:23:23', 'YYYY-MM-DD HH24:MI:SS')",
+    };
+    String tableName = getTableName();
+    tableNames.add(tableName);
+    createTableWithColTypes(types, vals);
+    // Some version of Oracle's jdbc drivers automatically convert date to
+    // timestamp. Since we don't want this to happen for this test,
+    // we must explicitly use a property file to control this behavior.
+    String connPropsFileName = "connection.properties";
+    FileUtils.writeStringToFile(new File(connPropsFileName), "oracle.jdbc.mapDateToTimestamp=false");
+    String[] args = getArgv(tableName, connPropsFileName, getColName(2));
+    runImport(args);
+
+    File file;
+    List<String> lines;
+
+    // First row should be in the first file
+    file = new File(this.getWarehouseDir(), "part-m-00000");
+    lines = FileUtils.readLines(file, "UTF-8");
+    assertEquals(1, lines.size());
+    assertEquals("1,old_data,1999-01-01 11:11:11.0", lines.get(0));
+
+    // With second line in the second file
+    file = new File(this.getWarehouseDir(), "part-m-00001");
+    lines = FileUtils.readLines(file, "UTF-8");
+    assertEquals(1, lines.size());
+    assertEquals("2,new_data,2000-11-11 23:23:23.0", lines.get(0));
+  }
+}


Mime
View raw message