flink-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From sunny patel <sunnyle...@gmail.com>
Subject java.lang.IllegalArgumentException: JDBC-Class not found. - org.postgresql.jdbc.Driver
Date Wed, 12 Oct 2016 21:18:27 GMT
Hi Guys,

I am facing JDBC error, could you please some one advise me on this error?

$ java -version

java version "1.8.0_102"

Java(TM) SE Runtime Environment (build 1.8.0_102-b14)

Java HotSpot(TM) 64-Bit Server VM (build 25.102-b14, mixed mode)

$ scala -version

Scala code runner version 2.11.8 -- Copyright 2002-2016, LAMP/EPFL


=============== Scala Code

import org.apache.flink.api.common.typeinfo.TypeInformation
import org.apache.flink.api.java.io.jdbc.JDBCInputFormat
import org.apache.flink.api.scala._
import org.apache.flink.api.table.typeutils.RowTypeInfo

object WordCount {
  def main(args: Array[String]) {

    val PATH = getClass.getResource("").getPath

    // set up the execution environment
    val env = ExecutionEnvironment.getExecutionEnvironment

    // Read data from JDBC (Kylin in our case)
    val stringColum: TypeInformation[Int] = createTypeInformation[Int]
    val DB_ROWTYPE = new RowTypeInfo(Seq(stringColum))

    val inputFormat = JDBCInputFormat.buildJDBCInputFormat()
      .setDrivername("org.postgresql.jdbc.Driver")
      .setDBUrl("jdbc:postgresql://localhost:5432/mydb")
      .setUsername("MI")
      .setPassword("MI")
      .setQuery("select * FROM identity")
      .setRowTypeInfo(DB_ROWTYPE)
      .finish()

    val dataset =env.createInput(inputFormat)
    dataset.print()

    println(PATH)
  }
}

==========================================================================

==========POM.XML


<?xml version="1.0" encoding="UTF-8"?>
<project xmlns="http://maven.apache.org/POM/4.0.0"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0
http://maven.apache.org/xsd/maven-4.0.0.xsd">
   <modelVersion>4.0.0</modelVersion>
   <parent>
      <artifactId>flink-parent</artifactId>
      <groupId>org.apache.flink</groupId>
      <version>1.2-SNAPSHOT</version>
   </parent>

   <groupId>org.apache.flink.quickstart</groupId>
   <artifactId>flink-scala-project</artifactId>
   <version>0.1</version>
   <packaging>jar</packaging>

   <name>Flink Quickstart Job</name>
   <url>http://www.myorganization.org</url>

   <repositories>
      <repository>
         <id>apache.snapshots</id>
         <name>Apache Development Snapshot Repository</name>
         <url>https://repository.apache.org/content/repositories/snapshots/</url>
         <releases>
            <enabled>false</enabled>
         </releases>
         <snapshots>
         </snapshots>
      </repository>
   </repositories>

   <properties>
      <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
      <flink.version>1.1.2</flink.version>
   </properties>

   <!--

       Execute "mvn clean package -Pbuild-jar"
       to build a jar file out of this project!

       How to use the Flink Quickstart pom:

       a) Adding new dependencies:
          You can add dependencies to the list below.
          Please check if the maven-shade-plugin below is filtering
out your dependency
          and remove the exclude from there.

       b) Build a jar for running on the cluster:
          There are two options for creating a jar from this project

          b.1) "mvn clean package" -> this will create a fat jar which
contains all
                dependencies necessary for running the jar created by
this pom in a cluster.
                The "maven-shade-plugin" excludes everything that is
provided on a running Flink cluster.

          b.2) "mvn clean package -Pbuild-jar" -> This will also
create a fat-jar, but with much
                nicer dependency exclusion handling. This approach is
preferred and leads to
                much cleaner jar files.
    -->

   <dependencies>
      <dependency>
         <groupId>org.apache.flink</groupId>
         <artifactId>flink-jdbc</artifactId>
         <version>${flink.version}</version>
      </dependency>
      <dependency>
         <groupId>org.apache.flink</groupId>
         <artifactId>flink-table_2.11</artifactId>
         <version>${flink.version}</version>
      </dependency>
      <dependency>
         <groupId>org.apache.flink</groupId>
         <artifactId>flink-scala_2.11</artifactId>
         <version>${flink.version}</version>
      </dependency>
      <dependency>
         <groupId>org.apache.flink</groupId>
         <artifactId>flink-streaming-scala_2.11</artifactId>
         <version>${flink.version}</version>
      </dependency>
      <dependency>
         <groupId>org.apache.flink</groupId>
         <artifactId>flink-clients_2.11</artifactId>
         <version>${flink.version}</version>
      </dependency>
   </dependencies>

   <profiles>
      <profile>
         <!-- Profile for packaging correct JAR files -->
         <id>build-jar</id>
         <activation>
         </activation>
         <dependencies>
            <dependency>
               <groupId>org.apache.flink</groupId>
               <artifactId>flink-scala_2.11</artifactId>
               <version>${flink.version}</version>
               <scope>provided</scope>
            </dependency>
            <dependency>
               <groupId>org.apache.flink</groupId>
               <artifactId>flink-streaming-scala_2.11</artifactId>
               <version>${flink.version}</version>
               <scope>provided</scope>
            </dependency>
            <dependency>
               <groupId>org.apache.flink</groupId>
               <artifactId>flink-clients_2.11</artifactId>
               <version>${flink.version}</version>
               <scope>provided</scope>
            </dependency>
         </dependencies>

         <build>
            <plugins>
               <!-- disable the exclusion rules -->
               <plugin>
                  <groupId>org.apache.maven.plugins</groupId>
                  <artifactId>maven-shade-plugin</artifactId>
                  <version>2.4.1</version>
                  <executions>
                     <execution>
                        <phase>package</phase>
                        <goals>
                           <goal>shade</goal>
                        </goals>
                        <configuration>
                           <artifactSet>
                              <excludes combine.self="override"></excludes>
                           </artifactSet>
                        </configuration>
                     </execution>
                  </executions>
               </plugin>
            </plugins>
         </build>
      </profile>
   </profiles>

   <!-- We use the maven-assembly plugin to create a fat jar that
contains all dependencies
       except flink and its transitive dependencies. The resulting
fat-jar can be executed
       on a cluster. Change the value of Program-Class if your program
entry point changes. -->
   <build>
      <plugins>
         <!-- We use the maven-shade plugin to create a fat jar that
contains all dependencies
            except flink and it's transitive dependencies. The
resulting fat-jar can be executed
            on a cluster. Change the value of Program-Class if your
program entry point changes. -->
         <plugin>
            <groupId>org.apache.maven.plugins</groupId>
            <artifactId>maven-shade-plugin</artifactId>
            <version>2.4.1</version>
            <executions>
               <!-- Run shade goal on package phase -->
               <execution>
                  <phase>package</phase>
                  <goals>
                     <goal>shade</goal>
                  </goals>
                  <configuration>
                     <artifactSet>
                        <excludes>
                           <!-- This list contains all dependencies of
flink-dist
                                    Everything else will be packaged
into the fat-jar
                                    -->
                           <exclude>org.apache.flink:flink-annotations</exclude>

<exclude>org.apache.flink:flink-shaded-hadoop1_2.11</exclude>

<exclude>org.apache.flink:flink-shaded-hadoop2</exclude>

<exclude>org.apache.flink:flink-shaded-curator-recipes</exclude>
                           <exclude>org.apache.flink:flink-core</exclude>
                           <exclude>org.apache.flink:flink-java</exclude>
                           <exclude>org.apache.flink:flink-scala_2.11</exclude>

<exclude>org.apache.flink:flink-runtime_2.11</exclude>

<exclude>org.apache.flink:flink-optimizer_2.11</exclude>

<exclude>org.apache.flink:flink-clients_2.11</exclude>
                           <exclude>org.apache.flink:flink-avro_2.11</exclude>

<exclude>org.apache.flink:flink-examples-batch_2.11</exclude>

<exclude>org.apache.flink:flink-examples-streaming_2.11</exclude>

<exclude>org.apache.flink:flink-streaming-java_2.11</exclude>

                           <!-- Also exclude very big transitive
dependencies of Flink

                                    WARNING: You have to remove these
excludes if your code relies on other
                                    versions of these dependencies.

                                    -->

                           <exclude>org.scala-lang:scala-library</exclude>
                           <exclude>org.scala-lang:scala-compiler</exclude>
                           <exclude>org.scala-lang:scala-reflect</exclude>
                           <exclude>com.typesafe.akka:akka-actor_*</exclude>
                           <exclude>com.typesafe.akka:akka-remote_*</exclude>
                           <exclude>com.typesafe.akka:akka-slf4j_*</exclude>
                           <exclude>io.netty:netty-all</exclude>
                           <exclude>io.netty:netty</exclude>

<exclude>commons-fileupload:commons-fileupload</exclude>
                           <exclude>org.apache.avro:avro</exclude>

<exclude>commons-collections:commons-collections</exclude>

<exclude>com.thoughtworks.paranamer:paranamer</exclude>
                           <exclude>org.xerial.snappy:snappy-java</exclude>

<exclude>org.apache.commons:commons-compress</exclude>
                           <exclude>org.tukaani:xz</exclude>
                           <exclude>com.esotericsoftware.kryo:kryo</exclude>
                           <exclude>com.esotericsoftware.minlog:minlog</exclude>
                           <exclude>org.objenesis:objenesis</exclude>
                           <exclude>com.twitter:chill_*</exclude>
                           <exclude>com.twitter:chill-java</exclude>
                           <exclude>commons-lang:commons-lang</exclude>
                           <exclude>junit:junit</exclude>
                           <exclude>org.apache.commons:commons-lang3</exclude>
                           <exclude>org.slf4j:slf4j-api</exclude>
                           <exclude>org.slf4j:slf4j-log4j12</exclude>
                           <exclude>log4j:log4j</exclude>
                           <exclude>org.apache.commons:commons-math</exclude>

<exclude>org.apache.sling:org.apache.sling.commons.json</exclude>
                           <exclude>commons-logging:commons-logging</exclude>
                           <exclude>commons-codec:commons-codec</exclude>

<exclude>com.fasterxml.jackson.core:jackson-core</exclude>

<exclude>com.fasterxml.jackson.core:jackson-databind</exclude>

<exclude>com.fasterxml.jackson.core:jackson-annotations</exclude>
                           <exclude>stax:stax-api</exclude>
                           <exclude>com.typesafe:config</exclude>

<exclude>org.uncommons.maths:uncommons-maths</exclude>
                           <exclude>com.github.scopt:scopt_*</exclude>
                           <exclude>commons-io:commons-io</exclude>
                           <exclude>commons-cli:commons-cli</exclude>
                        </excludes>
                     </artifactSet>
                     <filters>
                        <filter>
                           <artifact>org.apache.flink:*</artifact>
                           <excludes>
                              <!-- exclude shaded google but include
shaded curator -->
                              <exclude>org/apache/flink/shaded/com/**</exclude>
                              <exclude>web-docs/**</exclude>
                           </excludes>
                        </filter>
                        <filter>
                           <!-- Do not copy the signatures in the
META-INF folder.
                                    Otherwise, this might cause
SecurityExceptions when using the JAR. -->
                           <artifact>*:*</artifact>
                           <excludes>
                              <exclude>META-INF/*.SF</exclude>
                              <exclude>META-INF/*.DSA</exclude>
                              <exclude>META-INF/*.RSA</exclude>
                           </excludes>
                        </filter>
                     </filters>
                     <!-- If you want to use ./bin/flink run
<quickstart jar> uncomment the following lines.
                            This will add a Main-Class entry to the
manifest file -->
                     <!--
                            <transformers>
                               <transformer
implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">

<mainClass>org.apache.flink.quickstart.StreamingJob</mainClass>
                               </transformer>
                            </transformers>
                             -->

<createDependencyReducedPom>false</createDependencyReducedPom>
                  </configuration>
               </execution>
            </executions>
         </plugin>

         <plugin>
            <groupId>org.apache.maven.plugins</groupId>
            <artifactId>maven-compiler-plugin</artifactId>
            <version>3.1</version>
            <configuration>
               <source>1.7</source>
               <target>1.7</target>
            </configuration>
         </plugin>
         <plugin>
            <groupId>net.alchim31.maven</groupId>
            <artifactId>scala-maven-plugin</artifactId>
            <version>3.1.4</version>
            <executions>
               <execution>
                  <goals>
                     <goal>compile</goal>
                     <goal>testCompile</goal>
                  </goals>
               </execution>
            </executions>
         </plugin>

         <!-- Eclipse Integration -->
         <plugin>
            <groupId>org.apache.maven.plugins</groupId>
            <artifactId>maven-eclipse-plugin</artifactId>
            <version>2.8</version>
            <configuration>
               <downloadSources>true</downloadSources>
               <projectnatures>

<projectnature>org.scala-ide.sdt.core.scalanature</projectnature>
                  <projectnature>org.eclipse.jdt.core.javanature</projectnature>
               </projectnatures>
               <buildcommands>

<buildcommand>org.scala-ide.sdt.core.scalabuilder</buildcommand>
               </buildcommands>
               <classpathContainers>

<classpathContainer>org.scala-ide.sdt.launching.SCALA_CONTAINER</classpathContainer>

<classpathContainer>org.eclipse.jdt.launching.JRE_CONTAINER</classpathContainer>
               </classpathContainers>
               <excludes>
                  <exclude>org.scala-lang:scala-library</exclude>
                  <exclude>org.scala-lang:scala-compiler</exclude>
               </excludes>
               <sourceIncludes>
                  <sourceInclude>**/*.scala</sourceInclude>
                  <sourceInclude>**/*.java</sourceInclude>
               </sourceIncludes>
            </configuration>
         </plugin>

         <!-- Adding scala source directories to build path -->
         <plugin>
            <groupId>org.codehaus.mojo</groupId>
            <artifactId>build-helper-maven-plugin</artifactId>
            <version>1.7</version>
            <executions>
               <!-- Add src/main/scala to eclipse build path -->
               <execution>
                  <id>add-source</id>
                  <phase>generate-sources</phase>
                  <goals>
                     <goal>add-source</goal>
                  </goals>
                  <configuration>
                     <sources>
                        <source>src/main/scala</source>
                     </sources>
                  </configuration>
               </execution>
               <!-- Add src/test/scala to eclipse build path -->
               <execution>
                  <id>add-test-source</id>
                  <phase>generate-test-sources</phase>
                  <goals>
                     <goal>add-test-source</goal>
                  </goals>
                  <configuration>
                     <sources>
                        <source>src/test/scala</source>
                     </sources>
                  </configuration>
               </execution>
            </executions>
         </plugin>
      </plugins>
   </build>
</project>


==========================================================================

====== ERROR MESSAGE


/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/bin/java
-Didea.launcher.port=7532 "-Didea.launcher.bin.path=/Applications/IntelliJ
IDEA.app/Contents/bin" -Dfile.encoding=UTF-8 -classpath
"/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/deploy.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/cldrdata.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/dnsns.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/jaccess.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/jfxrt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/localedata.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/nashorn.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/sunec.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/sunjce_provider.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/sunpkcs11.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/ext/zipfs.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/javaws.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/jce.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/jfxswt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/jsse.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/management-agent.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/plugin.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/resources.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/rt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/ant-javafx.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/dt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/javafx-mx.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/jconsole.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/packager.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/sa-jdi.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/tools.jar:/Users/janaidu/MIPROJ/miaid/target/classes:/Users/janaidu/.m2/repository/org/apache/flink/flink-jdbc/1.1.2/flink-jdbc-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-table_2.11/1.1.2/flink-table_2.11-1.1.2.jar:/Users/janaidu/.m2/repository/org/codehaus/janino/janino/2.7.5/janino-2.7.5.jar:/Users/janaidu/.m2/repository/org/codehaus/janino/commons-compiler/2.7.5/commons-compiler-2.7.5.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-scala_2.11/1.1.2/flink-scala_2.11-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-core/1.1.2/flink-core-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-annotations/1.1.2/flink-annotations-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-metrics-core/1.1.2/flink-metrics-core-1.1.2.jar:/Users/janaidu/.m2/repository/com/esotericsoftware/kryo/kryo/2.24.0/kryo-2.24.0.jar:/Users/janaidu/.m2/repository/com/esotericsoftware/minlog/minlog/1.2/minlog-1.2.jar:/Users/janaidu/.m2/repository/org/apache/avro/avro/1.7.6/avro-1.7.6.jar:/Users/janaidu/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13.jar:/Users/janaidu/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar:/Users/janaidu/.m2/repository/com/thoughtworks/paranamer/paranamer/2.3/paranamer-2.3.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-shaded-hadoop2/1.1.2/flink-shaded-hadoop2-1.1.2.jar:/Users/janaidu/.m2/repository/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/Users/janaidu/.m2/repository/commons-codec/commons-codec/1.4/commons-codec-1.4.jar:/Users/janaidu/.m2/repository/commons-io/commons-io/2.4/commons-io-2.4.jar:/Users/janaidu/.m2/repository/commons-net/commons-net/3.1/commons-net-3.1.jar:/Users/janaidu/.m2/repository/commons-collections/commons-collections/3.2.2/commons-collections-3.2.2.jar:/Users/janaidu/.m2/repository/javax/servlet/servlet-api/2.5/servlet-api-2.5.jar:/Users/janaidu/.m2/repository/org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar:/Users/janaidu/.m2/repository/com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar:/Users/janaidu/.m2/repository/commons-el/commons-el/1.0/commons-el-1.0.jar:/Users/janaidu/.m2/repository/commons-logging/commons-logging/1.1.3/commons-logging-1.1.3.jar:/Users/janaidu/.m2/repository/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.jar:/Users/janaidu/.m2/repository/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/Users/janaidu/.m2/repository/commons-configuration/commons-configuration/1.7/commons-configuration-1.7.jar:/Users/janaidu/.m2/repository/commons-digester/commons-digester/1.8.1/commons-digester-1.8.1.jar:/Users/janaidu/.m2/repository/org/xerial/snappy/snappy-java/1.0.5/snappy-java-1.0.5.jar:/Users/janaidu/.m2/repository/com/jcraft/jsch/0.1.42/jsch-0.1.42.jar:/Users/janaidu/.m2/repository/org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar:/Users/janaidu/.m2/repository/org/tukaani/xz/1.0/xz-1.0.jar:/Users/janaidu/.m2/repository/commons-beanutils/commons-beanutils-bean-collections/1.8.3/commons-beanutils-bean-collections-1.8.3.jar:/Users/janaidu/.m2/repository/commons-daemon/commons-daemon/1.0.13/commons-daemon-1.0.13.jar:/Users/janaidu/.m2/repository/javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/Users/janaidu/.m2/repository/javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar:/Users/janaidu/.m2/repository/javax/activation/activation/1.1/activation-1.1.jar:/Users/janaidu/.m2/repository/com/google/inject/guice/3.0/guice-3.0.jar:/Users/janaidu/.m2/repository/javax/inject/javax.inject/1/javax.inject-1.jar:/Users/janaidu/.m2/repository/aopalliance/aopalliance/1.0/aopalliance-1.0.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-java/1.1.2/flink-java-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/commons/commons-math3/3.5/commons-math3-3.5.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-optimizer_2.11/1.1.2/flink-optimizer_2.11-1.1.2.jar:/Users/janaidu/.m2/repository/org/scala-lang/scala-reflect/2.11.7/scala-reflect-2.11.7.jar:/Users/janaidu/.m2/repository/org/scala-lang/scala-library/2.11.7/scala-library-2.11.7.jar:/Users/janaidu/.m2/repository/org/scala-lang/scala-compiler/2.11.7/scala-compiler-2.11.7.jar:/Users/janaidu/.m2/repository/org/scala-lang/modules/scala-xml_2.11/1.0.4/scala-xml_2.11-1.0.4.jar:/Users/janaidu/.m2/repository/org/scala-lang/modules/scala-parser-combinators_2.11/1.0.4/scala-parser-combinators_2.11-1.0.4.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-streaming-scala_2.11/1.1.2/flink-streaming-scala_2.11-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-streaming-java_2.11/1.1.2/flink-streaming-java_2.11-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/sling/org.apache.sling.commons.json/2.0.6/org.apache.sling.commons.json-2.0.6.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-clients_2.11/1.1.2/flink-clients_2.11-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/flink/flink-runtime_2.11/1.1.2/flink-runtime_2.11-1.1.2.jar:/Users/janaidu/.m2/repository/io/netty/netty-all/4.0.27.Final/netty-all-4.0.27.Final.jar:/Users/janaidu/.m2/repository/org/javassist/javassist/3.18.2-GA/javassist-3.18.2-GA.jar:/Users/janaidu/.m2/repository/com/typesafe/akka/akka-actor_2.11/2.3.7/akka-actor_2.11-2.3.7.jar:/Users/janaidu/.m2/repository/com/typesafe/config/1.2.1/config-1.2.1.jar:/Users/janaidu/.m2/repository/com/typesafe/akka/akka-remote_2.11/2.3.7/akka-remote_2.11-2.3.7.jar:/Users/janaidu/.m2/repository/io/netty/netty/3.8.0.Final/netty-3.8.0.Final.jar:/Users/janaidu/.m2/repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar:/Users/janaidu/.m2/repository/org/uncommons/maths/uncommons-maths/1.2.2a/uncommons-maths-1.2.2a.jar:/Users/janaidu/.m2/repository/com/typesafe/akka/akka-slf4j_2.11/2.3.7/akka-slf4j_2.11-2.3.7.jar:/Users/janaidu/.m2/repository/org/clapper/grizzled-slf4j_2.11/1.0.2/grizzled-slf4j_2.11-1.0.2.jar:/Users/janaidu/.m2/repository/com/github/scopt/scopt_2.11/3.2.0/scopt_2.11-3.2.0.jar:/Users/janaidu/.m2/repository/io/dropwizard/metrics/metrics-core/3.1.0/metrics-core-3.1.0.jar:/Users/janaidu/.m2/repository/io/dropwizard/metrics/metrics-jvm/3.1.0/metrics-jvm-3.1.0.jar:/Users/janaidu/.m2/repository/io/dropwizard/metrics/metrics-json/3.1.0/metrics-json-3.1.0.jar:/Users/janaidu/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.7.4/jackson-databind-2.7.4.jar:/Users/janaidu/.m2/repository/com/fasterxml/jackson/core/jackson-annotations/2.7.4/jackson-annotations-2.7.4.jar:/Users/janaidu/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.7.4/jackson-core-2.7.4.jar:/Users/janaidu/.m2/repository/org/apache/zookeeper/zookeeper/3.4.6/zookeeper-3.4.6.jar:/Users/janaidu/.m2/repository/jline/jline/0.9.94/jline-0.9.94.jar:/Users/janaidu/.m2/repository/com/twitter/chill_2.11/0.7.4/chill_2.11-0.7.4.jar:/Users/janaidu/.m2/repository/com/twitter/chill-java/0.7.4/chill-java-0.7.4.jar:/Users/janaidu/.m2/repository/commons-cli/commons-cli/1.3.1/commons-cli-1.3.1.jar:/Users/janaidu/.m2/repository/org/apache/flink/force-shading/1.2-SNAPSHOT/force-shading-1.2-20161012.043246-121.jar:/Users/janaidu/.m2/repository/com/google/code/findbugs/jsr305/1.3.9/jsr305-1.3.9.jar:/Users/janaidu/.m2/repository/org/apache/commons/commons-lang3/3.3.2/commons-lang3-3.3.2.jar:/Users/janaidu/.m2/repository/org/slf4j/slf4j-api/1.7.7/slf4j-api-1.7.7.jar:/Users/janaidu/.m2/repository/org/slf4j/slf4j-log4j12/1.7.7/slf4j-log4j12-1.7.7.jar:/Users/janaidu/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar:/Users/janaidu/.m2/repository/org/objenesis/objenesis/2.1/objenesis-2.1.jar:/Applications/IntelliJ
IDEA.app/Contents/lib/idea_rt.jar"
com.intellij.rt.execution.application.AppMain WordCount

log4j:WARN No appenders could be found for logger
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat).

log4j:WARN Please initialize the log4j system properly.

log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
more info.

Connected to JobManager at Actor[akka://flink/user/jobmanager_1#1408268854]

10/12/2016 20:41:59 Job execution switched to status RUNNING.

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(1/4) switched to
SCHEDULED

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(1/4) switched to
DEPLOYING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(2/4) switched to
SCHEDULED

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(2/4) switched to
DEPLOYING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
SCHEDULED

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
DEPLOYING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(4/4) switched to
SCHEDULED

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(4/4) switched to
DEPLOYING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(1/4) switched to
RUNNING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(4/4) switched to
RUNNING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
RUNNING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(2/4) switched to
RUNNING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(1/4) switched to
FAILED

java.lang.IllegalArgumentException: JDBC-Class not found. -
org.postgresql.jdbc.Driver

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:146)

at
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:110)

at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)

at java.lang.Thread.run(Thread.java:745)

Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver

at java.net.URLClassLoader.findClass(URLClassLoader.java:381)

at java.lang.ClassLoader.loadClass(ClassLoader.java:424)

at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)

at java.lang.ClassLoader.loadClass(ClassLoader.java:357)

at java.lang.Class.forName0(Native Method)

at java.lang.Class.forName(Class.java:264)

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:136)

... 3 more


10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(2/4) switched to
FAILED

java.lang.IllegalArgumentException: JDBC-Class not found. -
org.postgresql.jdbc.Driver

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:146)

at
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:110)

at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)

at java.lang.Thread.run(Thread.java:745)

Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver

at java.net.URLClassLoader.findClass(URLClassLoader.java:381)

at java.lang.ClassLoader.loadClass(ClassLoader.java:424)

at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)

at java.lang.ClassLoader.loadClass(ClassLoader.java:357)

at java.lang.Class.forName0(Native Method)

at java.lang.Class.forName(Class.java:264)

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:136)

... 3 more


10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(4/4) switched to
FAILED

java.lang.IllegalArgumentException: JDBC-Class not found. -
org.postgresql.jdbc.Driver

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:146)

at
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:110)

at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)

at java.lang.Thread.run(Thread.java:745)

Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver

at java.net.URLClassLoader.findClass(URLClassLoader.java:381)

at java.lang.ClassLoader.loadClass(ClassLoader.java:424)

at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)

at java.lang.ClassLoader.loadClass(ClassLoader.java:357)

at java.lang.Class.forName0(Native Method)

at java.lang.Class.forName(Class.java:264)

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:136)

... 3 more


10/12/2016 20:41:59 Job execution switched to status FAILING.

java.lang.IllegalArgumentException: JDBC-Class not found. -
org.postgresql.jdbc.Driver

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:146)

at
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:110)

at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)

at java.lang.Thread.run(Thread.java:745)

Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver

at java.net.URLClassLoader.findClass(URLClassLoader.java:381)

at java.lang.ClassLoader.loadClass(ClassLoader.java:424)

at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)

at java.lang.ClassLoader.loadClass(ClassLoader.java:357)

at java.lang.Class.forName0(Native Method)

at java.lang.Class.forName(Class.java:264)

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:136)

... 3 more

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
CANCELING

10/12/2016 20:41:59 DataSource (at
org.apache.flink.api.scala.ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395)
(org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
CANCELED

10/12/2016 20:41:59 DataSink (collect())(1/4) switched to CANCELED

10/12/2016 20:41:59 DataSink (collect())(2/4) switched to CANCELED

10/12/2016 20:41:59 DataSink (collect())(3/4) switched to CANCELED

10/12/2016 20:41:59 DataSink (collect())(4/4) switched to CANCELED

10/12/2016 20:41:59 Job execution switched to status FAILED.

Exception in thread "main"
org.apache.flink.runtime.client.JobExecutionException: Job execution failed.

at
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$mcV$sp(JobManager.scala:822)

at
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:768)

at
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:768)

at
scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)

at
scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)

at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41)

at
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:401)

at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)

at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253)

at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346)

at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)

at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

Caused by: java.lang.IllegalArgumentException: JDBC-Class not found. -
org.postgresql.jdbc.Driver

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:146)

at
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:110)

at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)

at java.lang.Thread.run(Thread.java:745)

Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver

at java.net.URLClassLoader.findClass(URLClassLoader.java:381)

at java.lang.ClassLoader.loadClass(ClassLoader.java:424)

at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)

at java.lang.ClassLoader.loadClass(ClassLoader.java:357)

at java.lang.Class.forName0(Native Method)

at java.lang.Class.forName(Class.java:264)

at
org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(JDBCInputFormat.java:136)

... 3 more


Process finished with exit code 1


============

Thanks

Sunny.

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message