Hi Sunny,

please avoid crossposting to all mailing lists.
The dev@f.a.o list is for issues related to the development of Flink not
the development of Flink applications.

The error message is actually quite descriptive. Flink does not find the
JDBC driver class.
You need to add it to the classpath for example by adding the corresponding
Maven dependency to your pom file.

Fabian


2016-10-12 23:18 GMT+02:00 sunny patel <sunnyle...@gmail.com>:

>
> Hi Guys,
>
> I am facing JDBC error, could you please some one advise me on this error?
>
> $ java -version
>
> java version "1.8.0_102"
>
> Java(TM) SE Runtime Environment (build 1.8.0_102-b14)
>
> Java HotSpot(TM) 64-Bit Server VM (build 25.102-b14, mixed mode)
>
> $ scala -version
>
> Scala code runner version 2.11.8 -- Copyright 2002-2016, LAMP/EPFL
>
>
> =============== Scala Code
>
> import org.apache.flink.api.common.typeinfo.TypeInformation
> import org.apache.flink.api.java.io.jdbc.JDBCInputFormat
> import org.apache.flink.api.scala._
> import org.apache.flink.api.table.typeutils.RowTypeInfo
>
> object WordCount {
>   def main(args: Array[String]) {
>
>     val PATH = getClass.getResource("").getPath
>
>     // set up the execution environment
>     val env = ExecutionEnvironment.getExecutionEnvironment
>
>     // Read data from JDBC (Kylin in our case)
>     val stringColum: TypeInformation[Int] = createTypeInformation[Int]
>     val DB_ROWTYPE = new RowTypeInfo(Seq(stringColum))
>
>     val inputFormat = JDBCInputFormat.buildJDBCInputFormat()
>       .setDrivername("org.postgresql.jdbc.Driver")
>       .setDBUrl("jdbc:postgresql://localhost:5432/mydb")
>       .setUsername("MI")
>       .setPassword("MI")
>       .setQuery("select * FROM identity")
>       .setRowTypeInfo(DB_ROWTYPE)
>       .finish()
>
>     val dataset =env.createInput(inputFormat)
>     dataset.print()
>
>     println(PATH)
>   }
> }
>
> ==========================================================================
>
> ==========POM.XML
>
>
> <?xml version="1.0" encoding="UTF-8"?>
> <project xmlns="http://maven.apache.org/POM/4.0.0"; 
> xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"; 
> xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
> http://maven.apache.org/xsd/maven-4.0.0.xsd";>
>    <modelVersion>4.0.0</modelVersion>
>    <parent>
>       <artifactId>flink-parent</artifactId>
>       <groupId>org.apache.flink</groupId>
>       <version>1.2-SNAPSHOT</version>
>    </parent>
>
>    <groupId>org.apache.flink.quickstart</groupId>
>    <artifactId>flink-scala-project</artifactId>
>    <version>0.1</version>
>    <packaging>jar</packaging>
>
>    <name>Flink Quickstart Job</name>
>    <url>http://www.myorganization.org</url>
>
>    <repositories>
>       <repository>
>          <id>apache.snapshots</id>
>          <name>Apache Development Snapshot Repository</name>
>          
> <url>https://repository.apache.org/content/repositories/snapshots/</url>
>          <releases>
>             <enabled>false</enabled>
>          </releases>
>          <snapshots>
>          </snapshots>
>       </repository>
>    </repositories>
>
>    <properties>
>       <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
>       <flink.version>1.1.2</flink.version>
>    </properties>
>
>    <!--
>
>        Execute "mvn clean package -Pbuild-jar"
>        to build a jar file out of this project!
>
>        How to use the Flink Quickstart pom:
>
>        a) Adding new dependencies:
>           You can add dependencies to the list below.
>           Please check if the maven-shade-plugin below is filtering out your 
> dependency
>           and remove the exclude from there.
>
>        b) Build a jar for running on the cluster:
>           There are two options for creating a jar from this project
>
>           b.1) "mvn clean package" -> this will create a fat jar which 
> contains all
>                 dependencies necessary for running the jar created by this 
> pom in a cluster.
>                 The "maven-shade-plugin" excludes everything that is provided 
> on a running Flink cluster.
>
>           b.2) "mvn clean package -Pbuild-jar" -> This will also create a 
> fat-jar, but with much
>                 nicer dependency exclusion handling. This approach is 
> preferred and leads to
>                 much cleaner jar files.
>     -->
>
>    <dependencies>
>       <dependency>
>          <groupId>org.apache.flink</groupId>
>          <artifactId>flink-jdbc</artifactId>
>          <version>${flink.version}</version>
>       </dependency>
>       <dependency>
>          <groupId>org.apache.flink</groupId>
>          <artifactId>flink-table_2.11</artifactId>
>          <version>${flink.version}</version>
>       </dependency>
>       <dependency>
>          <groupId>org.apache.flink</groupId>
>          <artifactId>flink-scala_2.11</artifactId>
>          <version>${flink.version}</version>
>       </dependency>
>       <dependency>
>          <groupId>org.apache.flink</groupId>
>          <artifactId>flink-streaming-scala_2.11</artifactId>
>          <version>${flink.version}</version>
>       </dependency>
>       <dependency>
>          <groupId>org.apache.flink</groupId>
>          <artifactId>flink-clients_2.11</artifactId>
>          <version>${flink.version}</version>
>       </dependency>
>    </dependencies>
>
>    <profiles>
>       <profile>
>          <!-- Profile for packaging correct JAR files -->
>          <id>build-jar</id>
>          <activation>
>          </activation>
>          <dependencies>
>             <dependency>
>                <groupId>org.apache.flink</groupId>
>                <artifactId>flink-scala_2.11</artifactId>
>                <version>${flink.version}</version>
>                <scope>provided</scope>
>             </dependency>
>             <dependency>
>                <groupId>org.apache.flink</groupId>
>                <artifactId>flink-streaming-scala_2.11</artifactId>
>                <version>${flink.version}</version>
>                <scope>provided</scope>
>             </dependency>
>             <dependency>
>                <groupId>org.apache.flink</groupId>
>                <artifactId>flink-clients_2.11</artifactId>
>                <version>${flink.version}</version>
>                <scope>provided</scope>
>             </dependency>
>          </dependencies>
>
>          <build>
>             <plugins>
>                <!-- disable the exclusion rules -->
>                <plugin>
>                   <groupId>org.apache.maven.plugins</groupId>
>                   <artifactId>maven-shade-plugin</artifactId>
>                   <version>2.4.1</version>
>                   <executions>
>                      <execution>
>                         <phase>package</phase>
>                         <goals>
>                            <goal>shade</goal>
>                         </goals>
>                         <configuration>
>                            <artifactSet>
>                               <excludes combine.self="override"></excludes>
>                            </artifactSet>
>                         </configuration>
>                      </execution>
>                   </executions>
>                </plugin>
>             </plugins>
>          </build>
>       </profile>
>    </profiles>
>
>    <!-- We use the maven-assembly plugin to create a fat jar that contains 
> all dependencies
>        except flink and its transitive dependencies. The resulting fat-jar 
> can be executed
>        on a cluster. Change the value of Program-Class if your program entry 
> point changes. -->
>    <build>
>       <plugins>
>          <!-- We use the maven-shade plugin to create a fat jar that contains 
> all dependencies
>             except flink and it's transitive dependencies. The resulting 
> fat-jar can be executed
>             on a cluster. Change the value of Program-Class if your program 
> entry point changes. -->
>          <plugin>
>             <groupId>org.apache.maven.plugins</groupId>
>             <artifactId>maven-shade-plugin</artifactId>
>             <version>2.4.1</version>
>             <executions>
>                <!-- Run shade goal on package phase -->
>                <execution>
>                   <phase>package</phase>
>                   <goals>
>                      <goal>shade</goal>
>                   </goals>
>                   <configuration>
>                      <artifactSet>
>                         <excludes>
>                            <!-- This list contains all dependencies of 
> flink-dist
>                                     Everything else will be packaged into the 
> fat-jar
>                                     -->
>                            
> <exclude>org.apache.flink:flink-annotations</exclude>
>                            
> <exclude>org.apache.flink:flink-shaded-hadoop1_2.11</exclude>
>                            
> <exclude>org.apache.flink:flink-shaded-hadoop2</exclude>
>                            
> <exclude>org.apache.flink:flink-shaded-curator-recipes</exclude>
>                            <exclude>org.apache.flink:flink-core</exclude>
>                            <exclude>org.apache.flink:flink-java</exclude>
>                            
> <exclude>org.apache.flink:flink-scala_2.11</exclude>
>                            
> <exclude>org.apache.flink:flink-runtime_2.11</exclude>
>                            
> <exclude>org.apache.flink:flink-optimizer_2.11</exclude>
>                            
> <exclude>org.apache.flink:flink-clients_2.11</exclude>
>                            <exclude>org.apache.flink:flink-avro_2.11</exclude>
>                            
> <exclude>org.apache.flink:flink-examples-batch_2.11</exclude>
>                            
> <exclude>org.apache.flink:flink-examples-streaming_2.11</exclude>
>                            
> <exclude>org.apache.flink:flink-streaming-java_2.11</exclude>
>
>                            <!-- Also exclude very big transitive dependencies 
> of Flink
>
>                                     WARNING: You have to remove these 
> excludes if your code relies on other
>                                     versions of these dependencies.
>
>                                     -->
>
>                            <exclude>org.scala-lang:scala-library</exclude>
>                            <exclude>org.scala-lang:scala-compiler</exclude>
>                            <exclude>org.scala-lang:scala-reflect</exclude>
>                            <exclude>com.typesafe.akka:akka-actor_*</exclude>
>                            <exclude>com.typesafe.akka:akka-remote_*</exclude>
>                            <exclude>com.typesafe.akka:akka-slf4j_*</exclude>
>                            <exclude>io.netty:netty-all</exclude>
>                            <exclude>io.netty:netty</exclude>
>                            
> <exclude>commons-fileupload:commons-fileupload</exclude>
>                            <exclude>org.apache.avro:avro</exclude>
>                            
> <exclude>commons-collections:commons-collections</exclude>
>                            
> <exclude>com.thoughtworks.paranamer:paranamer</exclude>
>                            <exclude>org.xerial.snappy:snappy-java</exclude>
>                            
> <exclude>org.apache.commons:commons-compress</exclude>
>                            <exclude>org.tukaani:xz</exclude>
>                            <exclude>com.esotericsoftware.kryo:kryo</exclude>
>                            
> <exclude>com.esotericsoftware.minlog:minlog</exclude>
>                            <exclude>org.objenesis:objenesis</exclude>
>                            <exclude>com.twitter:chill_*</exclude>
>                            <exclude>com.twitter:chill-java</exclude>
>                            <exclude>commons-lang:commons-lang</exclude>
>                            <exclude>junit:junit</exclude>
>                            <exclude>org.apache.commons:commons-lang3</exclude>
>                            <exclude>org.slf4j:slf4j-api</exclude>
>                            <exclude>org.slf4j:slf4j-log4j12</exclude>
>                            <exclude>log4j:log4j</exclude>
>                            <exclude>org.apache.commons:commons-math</exclude>
>                            
> <exclude>org.apache.sling:org.apache.sling.commons.json</exclude>
>                            <exclude>commons-logging:commons-logging</exclude>
>                            <exclude>commons-codec:commons-codec</exclude>
>                            
> <exclude>com.fasterxml.jackson.core:jackson-core</exclude>
>                            
> <exclude>com.fasterxml.jackson.core:jackson-databind</exclude>
>                            
> <exclude>com.fasterxml.jackson.core:jackson-annotations</exclude>
>                            <exclude>stax:stax-api</exclude>
>                            <exclude>com.typesafe:config</exclude>
>                            
> <exclude>org.uncommons.maths:uncommons-maths</exclude>
>                            <exclude>com.github.scopt:scopt_*</exclude>
>                            <exclude>commons-io:commons-io</exclude>
>                            <exclude>commons-cli:commons-cli</exclude>
>                         </excludes>
>                      </artifactSet>
>                      <filters>
>                         <filter>
>                            <artifact>org.apache.flink:*</artifact>
>                            <excludes>
>                               <!-- exclude shaded google but include shaded 
> curator -->
>                               
> <exclude>org/apache/flink/shaded/com/**</exclude>
>                               <exclude>web-docs/**</exclude>
>                            </excludes>
>                         </filter>
>                         <filter>
>                            <!-- Do not copy the signatures in the META-INF 
> folder.
>                                     Otherwise, this might cause 
> SecurityExceptions when using the JAR. -->
>                            <artifact>*:*</artifact>
>                            <excludes>
>                               <exclude>META-INF/*.SF</exclude>
>                               <exclude>META-INF/*.DSA</exclude>
>                               <exclude>META-INF/*.RSA</exclude>
>                            </excludes>
>                         </filter>
>                      </filters>
>                      <!-- If you want to use ./bin/flink run <quickstart jar> 
> uncomment the following lines.
>                             This will add a Main-Class entry to the manifest 
> file -->
>                      <!--
>                             <transformers>
>                                <transformer 
> implementation="org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
>                                   
> <mainClass>org.apache.flink.quickstart.StreamingJob</mainClass>
>                                </transformer>
>                             </transformers>
>                              -->
>                      
> <createDependencyReducedPom>false</createDependencyReducedPom>
>                   </configuration>
>                </execution>
>             </executions>
>          </plugin>
>
>          <plugin>
>             <groupId>org.apache.maven.plugins</groupId>
>             <artifactId>maven-compiler-plugin</artifactId>
>             <version>3.1</version>
>             <configuration>
>                <source>1.7</source>
>                <target>1.7</target>
>             </configuration>
>          </plugin>
>          <plugin>
>             <groupId>net.alchim31.maven</groupId>
>             <artifactId>scala-maven-plugin</artifactId>
>             <version>3.1.4</version>
>             <executions>
>                <execution>
>                   <goals>
>                      <goal>compile</goal>
>                      <goal>testCompile</goal>
>                   </goals>
>                </execution>
>             </executions>
>          </plugin>
>
>          <!-- Eclipse Integration -->
>          <plugin>
>             <groupId>org.apache.maven.plugins</groupId>
>             <artifactId>maven-eclipse-plugin</artifactId>
>             <version>2.8</version>
>             <configuration>
>                <downloadSources>true</downloadSources>
>                <projectnatures>
>                   
> <projectnature>org.scala-ide.sdt.core.scalanature</projectnature>
>                   
> <projectnature>org.eclipse.jdt.core.javanature</projectnature>
>                </projectnatures>
>                <buildcommands>
>                   
> <buildcommand>org.scala-ide.sdt.core.scalabuilder</buildcommand>
>                </buildcommands>
>                <classpathContainers>
>                   
> <classpathContainer>org.scala-ide.sdt.launching.SCALA_CONTAINER</classpathContainer>
>                   
> <classpathContainer>org.eclipse.jdt.launching.JRE_CONTAINER</classpathContainer>
>                </classpathContainers>
>                <excludes>
>                   <exclude>org.scala-lang:scala-library</exclude>
>                   <exclude>org.scala-lang:scala-compiler</exclude>
>                </excludes>
>                <sourceIncludes>
>                   <sourceInclude>**/*.scala</sourceInclude>
>                   <sourceInclude>**/*.java</sourceInclude>
>                </sourceIncludes>
>             </configuration>
>          </plugin>
>
>          <!-- Adding scala source directories to build path -->
>          <plugin>
>             <groupId>org.codehaus.mojo</groupId>
>             <artifactId>build-helper-maven-plugin</artifactId>
>             <version>1.7</version>
>             <executions>
>                <!-- Add src/main/scala to eclipse build path -->
>                <execution>
>                   <id>add-source</id>
>                   <phase>generate-sources</phase>
>                   <goals>
>                      <goal>add-source</goal>
>                   </goals>
>                   <configuration>
>                      <sources>
>                         <source>src/main/scala</source>
>                      </sources>
>                   </configuration>
>                </execution>
>                <!-- Add src/test/scala to eclipse build path -->
>                <execution>
>                   <id>add-test-source</id>
>                   <phase>generate-test-sources</phase>
>                   <goals>
>                      <goal>add-test-source</goal>
>                   </goals>
>                   <configuration>
>                      <sources>
>                         <source>src/test/scala</source>
>                      </sources>
>                   </configuration>
>                </execution>
>             </executions>
>          </plugin>
>       </plugins>
>    </build>
> </project>
>
>
> ==========================================================================
>
> ====== ERROR MESSAGE
>
>
> /Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/bin/java
> -Didea.launcher.port=7532 "-Didea.launcher.bin.path=/Applications/IntelliJ
> IDEA.app/Contents/bin" -Dfile.encoding=UTF-8 -classpath "/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/
> charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/jre/lib/deploy.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/
> ext/cldrdata.jar:/Library/Java/JavaVirtualMachines/jdk1.
> 8.0_102.jdk/Contents/Home/jre/lib/ext/dnsns.jar:/Library/
> Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/
> lib/ext/jaccess.jar:/Library/Java/JavaVirtualMachines/jdk1.
> 8.0_102.jdk/Contents/Home/jre/lib/ext/jfxrt.jar:/Library/
> Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/
> lib/ext/localedata.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/jre/lib/ext/nashorn.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/
> ext/sunec.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/jre/lib/ext/sunjce_provider.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/
> ext/sunpkcs11.jar:/Library/Java/JavaVirtualMachines/jdk1.
> 8.0_102.jdk/Contents/Home/jre/lib/ext/zipfs.jar:/Library/
> Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/
> lib/javaws.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/jre/lib/jce.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/
> jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/jre/lib/jfxswt.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/
> jsse.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/jre/lib/management-agent.jar:/Library/
> Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/
> lib/plugin.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/jre/lib/resources.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/jre/lib/
> rt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/lib/ant-javafx.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/dt.
> jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/
> javafx-mx.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/lib/jconsole.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/
> packager.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_
> 102.jdk/Contents/Home/lib/sa-jdi.jar:/Library/Java/
> JavaVirtualMachines/jdk1.8.0_102.jdk/Contents/Home/lib/
> tools.jar:/Users/janaidu/MIPROJ/miaid/target/classes:/
> Users/janaidu/.m2/repository/org/apache/flink/flink-jdbc/1.
> 1.2/flink-jdbc-1.1.2.jar:/Users/janaidu/.m2/repository/
> org/apache/flink/flink-table_2.11/1.1.2/flink-table_2.11-1.
> 1.2.jar:/Users/janaidu/.m2/repository/org/codehaus/
> janino/janino/2.7.5/janino-2.7.5.jar:/Users/janaidu/.m2/
> repository/org/codehaus/janino/commons-compiler/2.7.5/
> commons-compiler-2.7.5.jar:/Users/janaidu/.m2/repository/
> org/apache/flink/flink-scala_2.11/1.1.2/flink-scala_2.11-1.
> 1.2.jar:/Users/janaidu/.m2/repository/org/apache/flink/
> flink-core/1.1.2/flink-core-1.1.2.jar:/Users/janaidu/.m2/
> repository/org/apache/flink/flink-annotations/1.1.2/flink-
> annotations-1.1.2.jar:/Users/janaidu/.m2/repository/org/
> apache/flink/flink-metrics-core/1.1.2/flink-metrics-core-
> 1.1.2.jar:/Users/janaidu/.m2/repository/com/esotericsoftware/kryo/kryo/2.
> 24.0/kryo-2.24.0.jar:/Users/janaidu/.m2/repository/com/
> esotericsoftware/minlog/minlog/1.2/minlog-1.2.jar:/
> Users/janaidu/.m2/repository/org/apache/avro/avro/1.7.6/
> avro-1.7.6.jar:/Users/janaidu/.m2/repository/org/codehaus/
> jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13.
> jar:/Users/janaidu/.m2/repository/org/codehaus/
> jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.
> 13.jar:/Users/janaidu/.m2/repository/com/thoughtworks/
> paranamer/paranamer/2.3/paranamer-2.3.jar:/Users/
> janaidu/.m2/repository/org/apache/flink/flink-shaded-
> hadoop2/1.1.2/flink-shaded-hadoop2-1.1.2.jar:/Users/
> janaidu/.m2/repository/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/
> Users/janaidu/.m2/repository/commons-codec/commons-codec/1.
> 4/commons-codec-1.4.jar:/Users/janaidu/.m2/repository/
> commons-io/commons-io/2.4/commons-io-2.4.jar:/Users/
> janaidu/.m2/repository/commons-net/commons-net/3.1/
> commons-net-3.1.jar:/Users/janaidu/.m2/repository/
> commons-collections/commons-collections/3.2.2/commons-
> collections-3.2.2.jar:/Users/janaidu/.m2/repository/javax/
> servlet/servlet-api/2.5/servlet-api-2.5.jar:/Users/
> janaidu/.m2/repository/org/mortbay/jetty/jetty-util/6.1.
> 26/jetty-util-6.1.26.jar:/Users/janaidu/.m2/repository/
> com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar:/Users/
> janaidu/.m2/repository/commons-el/commons-el/1.0/
> commons-el-1.0.jar:/Users/janaidu/.m2/repository/commons-logging/commons-
> logging/1.1.3/commons-logging-1.1.3.jar:/Users/janaidu/.m2/
> repository/com/jamesmurty/utils/java-xmlbuilder/0.4/
> java-xmlbuilder-0.4.jar:/Users/janaidu/.m2/repository/
> commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/Users/
> janaidu/.m2/repository/commons-configuration/commons-
> configuration/1.7/commons-configuration-1.7.jar:/Users/
> janaidu/.m2/repository/commons-digester/commons-digester/1.8.1/commons-
> digester-1.8.1.jar:/Users/janaidu/.m2/repository/org/
> xerial/snappy/snappy-java/1.0.5/snappy-java-1.0.5.jar:/
> Users/janaidu/.m2/repository/com/jcraft/jsch/0.1.42/jsch-0.
> 1.42.jar:/Users/janaidu/.m2/repository/org/apache/commons/
> commons-compress/1.4.1/commons-compress-1.4.1.jar:/
> Users/janaidu/.m2/repository/org/tukaani/xz/1.0/xz-1.0.jar:
> /Users/janaidu/.m2/repository/commons-beanutils/commons-
> beanutils-bean-collections/1.8.3/commons-beanutils-bean-
> collections-1.8.3.jar:/Users/janaidu/.m2/repository/
> commons-daemon/commons-daemon/1.0.13/commons-daemon-1.0.13.
> jar:/Users/janaidu/.m2/repository/javax/xml/bind/
> jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/Users/janaidu/.m2/
> repository/javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.
> jar:/Users/janaidu/.m2/repository/javax/activation/
> activation/1.1/activation-1.1.jar:/Users/janaidu/.m2/
> repository/com/google/inject/guice/3.0/guice-3.0.jar:/
> Users/janaidu/.m2/repository/javax/inject/javax.inject/1/
> javax.inject-1.jar:/Users/janaidu/.m2/repository/
> aopalliance/aopalliance/1.0/aopalliance-1.0.jar:/Users/
> janaidu/.m2/repository/org/apache/flink/flink-java/1.1.2/
> flink-java-1.1.2.jar:/Users/janaidu/.m2/repository/org/
> apache/commons/commons-math3/3.5/commons-math3-3.5.jar:/
> Users/janaidu/.m2/repository/org/apache/flink/flink-
> optimizer_2.11/1.1.2/flink-optimizer_2.11-1.1.2.jar:/
> Users/janaidu/.m2/repository/org/scala-lang/scala-reflect/
> 2.11.7/scala-reflect-2.11.7.jar:/Users/janaidu/.m2/
> repository/org/scala-lang/scala-library/2.11.7/scala-
> library-2.11.7.jar:/Users/janaidu/.m2/repository/org/
> scala-lang/scala-compiler/2.11.7/scala-compiler-2.11.7.
> jar:/Users/janaidu/.m2/repository/org/scala-lang/
> modules/scala-xml_2.11/1.0.4/scala-xml_2.11-1.0.4.jar:/
> Users/janaidu/.m2/repository/org/scala-lang/modules/scala-
> parser-combinators_2.11/1.0.4/scala-parser-combinators_2.11-
> 1.0.4.jar:/Users/janaidu/.m2/repository/org/apache/flink/
> flink-streaming-scala_2.11/1.1.2/flink-streaming-scala_2.
> 11-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/
> flink/flink-streaming-java_2.11/1.1.2/flink-streaming-java_
> 2.11-1.1.2.jar:/Users/janaidu/.m2/repository/org/apache/
> sling/org.apache.sling.commons.json/2.0.6/org.apache.
> sling.commons.json-2.0.6.jar:/Users/janaidu/.m2/repository/
> org/apache/flink/flink-clients_2.11/1.1.2/flink-
> clients_2.11-1.1.2.jar:/Users/janaidu/.m2/repository/org/
> apache/flink/flink-runtime_2.11/1.1.2/flink-runtime_2.11-1.
> 1.2.jar:/Users/janaidu/.m2/repository/io/netty/netty-all/
> 4.0.27.Final/netty-all-4.0.27.Final.jar:/Users/janaidu/.m2/
> repository/org/javassist/javassist/3.18.2-GA/javassist-
> 3.18.2-GA.jar:/Users/janaidu/.m2/repository/com/typesafe/
> akka/akka-actor_2.11/2.3.7/akka-actor_2.11-2.3.7.jar:/
> Users/janaidu/.m2/repository/com/typesafe/config/1.2.1/
> config-1.2.1.jar:/Users/janaidu/.m2/repository/com/
> typesafe/akka/akka-remote_2.11/2.3.7/akka-remote_2.11-2.3.
> 7.jar:/Users/janaidu/.m2/repository/io/netty/netty/3.8.
> 0.Final/netty-3.8.0.Final.jar:/Users/janaidu/.m2/repository/
> com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.
> 0.jar:/Users/janaidu/.m2/repository/org/uncommons/
> maths/uncommons-maths/1.2.2a/uncommons-maths-1.2.2a.jar:/
> Users/janaidu/.m2/repository/com/typesafe/akka/akka-slf4j_
> 2.11/2.3.7/akka-slf4j_2.11-2.3.7.jar:/Users/janaidu/.m2/
> repository/org/clapper/grizzled-slf4j_2.11/1.0.2/
> grizzled-slf4j_2.11-1.0.2.jar:/Users/janaidu/.m2/repository/
> com/github/scopt/scopt_2.11/3.2.0/scopt_2.11-3.2.0.jar:/
> Users/janaidu/.m2/repository/io/dropwizard/metrics/metrics-
> core/3.1.0/metrics-core-3.1.0.jar:/Users/janaidu/.m2/
> repository/io/dropwizard/metrics/metrics-jvm/3.1.0/
> metrics-jvm-3.1.0.jar:/Users/janaidu/.m2/repository/io/
> dropwizard/metrics/metrics-json/3.1.0/metrics-json-3.1.0.
> jar:/Users/janaidu/.m2/repository/com/fasterxml/
> jackson/core/jackson-databind/2.7.4/jackson-databind-2.7.4.
> jar:/Users/janaidu/.m2/repository/com/fasterxml/jackson/core/jackson-
> annotations/2.7.4/jackson-annotations-2.7.4.jar:/Users/
> janaidu/.m2/repository/com/fasterxml/jackson/core/
> jackson-core/2.7.4/jackson-core-2.7.4.jar:/Users/janaidu/
> .m2/repository/org/apache/zookeeper/zookeeper/3.4.6/
> zookeeper-3.4.6.jar:/Users/janaidu/.m2/repository/jline/
> jline/0.9.94/jline-0.9.94.jar:/Users/janaidu/.m2/repository/
> com/twitter/chill_2.11/0.7.4/chill_2.11-0.7.4.jar:/Users/
> janaidu/.m2/repository/com/twitter/chill-java/0.7.4/
> chill-java-0.7.4.jar:/Users/janaidu/.m2/repository/
> commons-cli/commons-cli/1.3.1/commons-cli-1.3.1.jar:/Users/
> janaidu/.m2/repository/org/apache/flink/force-shading/1.
> 2-SNAPSHOT/force-shading-1.2-20161012.043246-121.jar:/
> Users/janaidu/.m2/repository/com/google/code/findbugs/
> jsr305/1.3.9/jsr305-1.3.9.jar:/Users/janaidu/.m2/repository/
> org/apache/commons/commons-lang3/3.3.2/commons-lang3-3.3.
> 2.jar:/Users/janaidu/.m2/repository/org/slf4j/slf4j-
> api/1.7.7/slf4j-api-1.7.7.jar:/Users/janaidu/.m2/repository/
> org/slf4j/slf4j-log4j12/1.7.7/slf4j-log4j12-1.7.7.jar:/
> Users/janaidu/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.
> 17.jar:/Users/janaidu/.m2/repository/org/objenesis/
> objenesis/2.1/objenesis-2.1.jar:/Applications/IntelliJ
> IDEA.app/Contents/lib/idea_rt.jar" 
> com.intellij.rt.execution.application.AppMain
> WordCount
>
> log4j:WARN No appenders could be found for logger (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat).
>
> log4j:WARN Please initialize the log4j system properly.
>
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for
> more info.
>
> Connected to JobManager at Actor[akka://flink/user/
> jobmanager_1#1408268854]
>
> 10/12/2016 20:41:59 Job execution switched to status RUNNING.
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(1/4) switched to
> SCHEDULED
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(1/4) switched to
> DEPLOYING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(2/4) switched to
> SCHEDULED
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(2/4) switched to
> DEPLOYING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
> SCHEDULED
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
> DEPLOYING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(4/4) switched to
> SCHEDULED
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(4/4) switched to
> DEPLOYING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(1/4) switched to
> RUNNING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(4/4) switched to
> RUNNING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
> RUNNING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(2/4) switched to
> RUNNING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(1/4) switched to
> FAILED
>
> java.lang.IllegalArgumentException: JDBC-Class not found. -
> org.postgresql.jdbc.Driver
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:146)
>
> at org.apache.flink.runtime.operators.DataSourceTask.
> invoke(DataSourceTask.java:110)
>
> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver
>
> at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>
> at java.lang.Class.forName0(Native Method)
>
> at java.lang.Class.forName(Class.java:264)
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:136)
>
> ... 3 more
>
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(2/4) switched to
> FAILED
>
> java.lang.IllegalArgumentException: JDBC-Class not found. -
> org.postgresql.jdbc.Driver
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:146)
>
> at org.apache.flink.runtime.operators.DataSourceTask.
> invoke(DataSourceTask.java:110)
>
> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver
>
> at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>
> at java.lang.Class.forName0(Native Method)
>
> at java.lang.Class.forName(Class.java:264)
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:136)
>
> ... 3 more
>
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(4/4) switched to
> FAILED
>
> java.lang.IllegalArgumentException: JDBC-Class not found. -
> org.postgresql.jdbc.Driver
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:146)
>
> at org.apache.flink.runtime.operators.DataSourceTask.
> invoke(DataSourceTask.java:110)
>
> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver
>
> at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>
> at java.lang.Class.forName0(Native Method)
>
> at java.lang.Class.forName(Class.java:264)
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:136)
>
> ... 3 more
>
>
> 10/12/2016 20:41:59 Job execution switched to status FAILING.
>
> java.lang.IllegalArgumentException: JDBC-Class not found. -
> org.postgresql.jdbc.Driver
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:146)
>
> at org.apache.flink.runtime.operators.DataSourceTask.
> invoke(DataSourceTask.java:110)
>
> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver
>
> at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>
> at java.lang.Class.forName0(Native Method)
>
> at java.lang.Class.forName(Class.java:264)
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:136)
>
> ... 3 more
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
> CANCELING
>
> 10/12/2016 20:41:59 DataSource (at org.apache.flink.api.scala.
> ExecutionEnvironment.createInput(ExecutionEnvironment.scala:395) (
> org.apache.flink.api.java.io.jdbc.JDBCInputFormat))(3/4) switched to
> CANCELED
>
> 10/12/2016 20:41:59 DataSink (collect())(1/4) switched to CANCELED
>
> 10/12/2016 20:41:59 DataSink (collect())(2/4) switched to CANCELED
>
> 10/12/2016 20:41:59 DataSink (collect())(3/4) switched to CANCELED
>
> 10/12/2016 20:41:59 DataSink (collect())(4/4) switched to CANCELED
>
> 10/12/2016 20:41:59 Job execution switched to status FAILED.
>
> Exception in thread "main" 
> org.apache.flink.runtime.client.JobExecutionException:
> Job execution failed.
>
> at org.apache.flink.runtime.jobmanager.JobManager$$
> anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply$
> mcV$sp(JobManager.scala:822)
>
> at org.apache.flink.runtime.jobmanager.JobManager$$
> anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:768)
>
> at org.apache.flink.runtime.jobmanager.JobManager$$
> anonfun$handleMessage$1$$anonfun$applyOrElse$7.apply(JobManager.scala:768)
>
> at scala.concurrent.impl.Future$PromiseCompletingRunnable.
> liftedTree1$1(Future.scala:24)
>
> at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(
> Future.scala:24)
>
> at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:41)
>
> at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(
> AbstractDispatcher.scala:401)
>
> at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>
> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.
> pollAndExecAll(ForkJoinPool.java:1253)
>
> at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.
> runTask(ForkJoinPool.java:1346)
>
> at scala.concurrent.forkjoin.ForkJoinPool.runWorker(
> ForkJoinPool.java:1979)
>
> at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(
> ForkJoinWorkerThread.java:107)
>
> Caused by: java.lang.IllegalArgumentException: JDBC-Class not found. -
> org.postgresql.jdbc.Driver
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:146)
>
> at org.apache.flink.runtime.operators.DataSourceTask.
> invoke(DataSourceTask.java:110)
>
> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:584)
>
> at java.lang.Thread.run(Thread.java:745)
>
> Caused by: java.lang.ClassNotFoundException: org.postgresql.jdbc.Driver
>
> at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
>
> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>
> at java.lang.Class.forName0(Native Method)
>
> at java.lang.Class.forName(Class.java:264)
>
> at org.apache.flink.api.java.io.jdbc.JDBCInputFormat.openInputFormat(
> JDBCInputFormat.java:136)
>
> ... 3 more
>
>
> Process finished with exit code 1
>
>
> ============
>
> Thanks
>
> Sunny.
>

Reply via email to