manifoldcf-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From kwri...@apache.org
Subject svn commit: r1784692 - in /manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka: APISanityHSQLDBIT.java BaseITHSQLDB.java
Date Tue, 28 Feb 2017 06:58:06 GMT
Author: kwright
Date: Tue Feb 28 06:58:06 2017
New Revision: 1784692

URL: http://svn.apache.org/viewvc?rev=1784692&view=rev
Log:
Get kafka APISanity test working

Modified:
    manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/APISanityHSQLDBIT.java
    manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/BaseITHSQLDB.java

Modified: manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/APISanityHSQLDBIT.java
URL: http://svn.apache.org/viewvc/manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/APISanityHSQLDBIT.java?rev=1784692&r1=1784691&r2=1784692&view=diff
==============================================================================
--- manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/APISanityHSQLDBIT.java
(original)
+++ manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/APISanityHSQLDBIT.java
Tue Feb 28 06:58:06 2017
@@ -31,7 +31,6 @@ public class APISanityHSQLDBIT extends B
   @Test
   public void sanityCheck()
           throws Exception {
-    try {
       int i;
 
       // Create a basic file system connection, and save it.
@@ -59,14 +58,6 @@ public class APISanityHSQLDBIT extends B
 
       child = new ConfigurationNode("configuration");
 
-      //Testing Repository Connector parameters
-      // MHL
-      /*
-       ConfigurationNode cmisBindingNode = new ConfigurationNode("_PARAMETER_");
-       cmisBindingNode.setAttribute("name", CmisConfig.BINDING_PARAM);
-       cmisBindingNode.setValue(CmisConfig.BINDING_DEFAULT_VALUE);
-       child.addChild(child.getChildCount(), cmisBindingNode);
-       */
       connectionObject.addChild(connectionObject.getChildCount(), child);
 
       requestObject = new Configuration();
@@ -199,10 +190,6 @@ public class APISanityHSQLDBIT extends B
       waitJobDeleted(jobIDString, 120000L);
 
       // Cleanup is automatic by the base class, so we can feel free to leave jobs and connections
lying around.
-    } catch (Exception e) {
-      e.printStackTrace();
-      throw e;
-    }
   }
 
   protected void startJob(String jobIDString)

Modified: manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/BaseITHSQLDB.java
URL: http://svn.apache.org/viewvc/manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/BaseITHSQLDB.java?rev=1784692&r1=1784691&r2=1784692&view=diff
==============================================================================
--- manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/BaseITHSQLDB.java
(original)
+++ manifoldcf/trunk/connectors/kafka/connector/src/test/java/org/apache/manifoldcf/agents/output/kafka/BaseITHSQLDB.java
Tue Feb 28 06:58:06 2017
@@ -17,6 +17,8 @@
 
 package org.apache.manifoldcf.agents.output.kafka;
 
+import java.io.File;
+
 import java.util.Properties;
 import org.junit.After;
 import static org.junit.Assert.fail;
@@ -48,41 +50,41 @@ public class BaseITHSQLDB extends org.ap
     Properties kafkaProperties = new Properties();
     Properties zkProperties = new Properties();
 
-    try {
-      //load properties
-      kafkaProperties.put("broker.id", "0");
-      kafkaProperties.put("port", "9092");
-      kafkaProperties.put("num.network.threads", "3");
-      kafkaProperties.put("num.io.threads", "8");
-      kafkaProperties.put("socket.send.buffer.bytes", "102400");
-      kafkaProperties.put("socket.receive.buffer.bytes", "102400");
-      kafkaProperties.put("socket.request.max.bytes", "104857600");
-      kafkaProperties.put("log.dirs", "/tmp/kafka-logs");
-      kafkaProperties.put("num.partitions", "1");
-      kafkaProperties.put("num.recovery.threads.per.data.dir", "1");
-      kafkaProperties.put("log.retention.hours", "168");
-      kafkaProperties.put("log.segment.bytes", "1073741824");
-      kafkaProperties.put("log.retention.check.interval.ms", "300000");
-      kafkaProperties.put("log.cleaner.enable", "false");
-      kafkaProperties.put("zookeeper.connect", "localhost:2181");
-      kafkaProperties.put("zookeeper.connection.timeout.ms", "6000");
-
-      zkProperties.put("dataDir", "/tmp/zookeeper");
-      zkProperties.put("clientPort", "2181");
-      zkProperties.put("maxClientCnxns", "0");
-
-      //kafkaProperties.load(Class.class.getResourceAsStream("/kafkalocal.properties"));
-      //zkProperties.load(Class.class.getResourceAsStream("/zklocal.properties"));
-      System.out.println("Kafka is starting...");
-
-      //start kafka
-      kafka = new KafkaLocal(kafkaProperties, zkProperties);
-      Thread.sleep(5000);
-    } catch (Exception e) {
-      e.printStackTrace(System.out);
-      fail("Error running local Kafka broker");
-      e.printStackTrace(System.out);
-    }
+    String tmpDir = System.getProperty("java.io.tmpdir");
+    File logDir = new File(tmpDir, "kafka-logs");
+    logDir.mkdir();
+    File zookeeperDir = new File(tmpDir, "zookeeper");
+    zookeeperDir.mkdir();
+            
+    //load properties
+    kafkaProperties.put("broker.id", "0");
+    kafkaProperties.put("port", "9092");
+    kafkaProperties.put("num.network.threads", "3");
+    kafkaProperties.put("num.io.threads", "8");
+    kafkaProperties.put("socket.send.buffer.bytes", "102400");
+    kafkaProperties.put("socket.receive.buffer.bytes", "102400");
+    kafkaProperties.put("socket.request.max.bytes", "104857600");
+    kafkaProperties.put("log.dirs", logDir.getAbsolutePath());
+    kafkaProperties.put("num.partitions", "1");
+    kafkaProperties.put("num.recovery.threads.per.data.dir", "1");
+    kafkaProperties.put("log.retention.hours", "168");
+    kafkaProperties.put("log.segment.bytes", "1073741824");
+    kafkaProperties.put("log.retention.check.interval.ms", "300000");
+    kafkaProperties.put("log.cleaner.enable", "false");
+    kafkaProperties.put("zookeeper.connect", "localhost:2181");
+    kafkaProperties.put("zookeeper.connection.timeout.ms", "6000");
+
+    zkProperties.put("dataDir", zookeeperDir.getAbsolutePath());
+    zkProperties.put("clientPort", "2181");
+    zkProperties.put("maxClientCnxns", "0");
+
+    //kafkaProperties.load(Class.class.getResourceAsStream("/kafkalocal.properties"));
+    //zkProperties.load(Class.class.getResourceAsStream("/zklocal.properties"));
+    System.out.println("Kafka is starting...");
+
+    //start kafka
+    kafka = new KafkaLocal(kafkaProperties, zkProperties);
+    Thread.sleep(5000);
   }
 
   @After



Mime
View raw message