Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Comment: add sbt tests

...

  • Run the start-local.sh/start-local-streaming.sh, start-cluster.sh/start-cluster-streaming.sh, start-webclient.sh scripts  scripts and verify that the processes come up
    • Examine the *.out files (should be empty) and the log files (should contain no exceptions)
    • Test for Linux, OS X, Windows (for Windows as far as possible, not all scripts exist)
    • Shutdown and verify there are no exceptions in the log output (after shutdown)
    • Check all start+submission scripts for paths with and without spaces (./bin/* scripts are quite fragile for paths with spaces)

...

  • Test on a cluster with HDFS.
    • Check that a good amount of input splits is read locally (JobManager log reveals local assignments)

  • Test against a Kafka installation

  • Test the ./bin/flink command line client
    • Test "info" option, paste the JSON into the plan visualizer HTML file, check that plan is rendered
    • Test the parallelism flag (-p) to override the configured default parallelism

  • Verify the plan visualizer with different browsers/operating systems

  • Verify that the quickstarts for scala and java are working with the staging repository for both IntelliJ and Eclipse.
    • in particular the dependencies of the quickstart project need to be set correctly and the QS project needs to build from the staging repository (replace the snapshot repo URL with the staging repo URL)
    • The dependency tree of the QuickStart project must not contain any dependencies we shade away upstream (guava, netty, ...)
    • Test that quickstart archetypes are working on all platforms

  • Run examples on a YARN cluster

  • Run all examples from the IDE (Eclipse & IntelliJ)

  • Run an example with the RemoteEnvironment against a cluster started from the shell script

  • Pay special attention to new features

  • Test recovery and exactly-once guarantees with master and worker failures @todo @uce Will update this with scripts
    • YARN (see https://github.com/apache/flink/pull/1213 for details)
      • 2.3.0 <= version < 2.4.0
        • Set yarn.application-attempts for Flink
        • Set yarn.resourcemanager.am.max-attempts for YARN (upper bound on number of failures)
        • Note: it's expected for these Hadoop versions that all containers are killed when the application master fails
      • 2.4.0 <= version < 2.6.0
        • Important: in this version the task manager containers should stay alive when the application master is killed
      • 2.6.0 <= version
        • Check that the application is only killed by YARN after the system has seen the maximum number of application attempts during one interval
    • Standalone
      • Start multiple JobManager and TaskManager instances
      • Kill random instances (make sure that enough task slots and standby job managers are available)
         
  • Test building a SBT project depending on Flink.

Creating a release candidate

...