flink-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From se...@apache.org
Subject [56/73] [abbrv] git commit: Rename documentation
Date Sat, 12 Jul 2014 12:48:35 GMT
Rename documentation


Project: http://git-wip-us.apache.org/repos/asf/incubator-flink/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-flink/commit/fbc93386
Tree: http://git-wip-us.apache.org/repos/asf/incubator-flink/tree/fbc93386
Diff: http://git-wip-us.apache.org/repos/asf/incubator-flink/diff/fbc93386

Branch: refs/heads/master
Commit: fbc93386645d28ab60c9935c27054fcfe224e861
Parents: ebe6b90
Author: Robert Metzger <rmetzger@apache.org>
Authored: Fri Jul 11 16:46:45 2014 +0200
Committer: Robert Metzger <rmetzger@apache.org>
Committed: Fri Jul 11 16:53:43 2014 +0200

----------------------------------------------------------------------
 docs/cli.md                     | 54 ++++++++++----------
 docs/cluster_execution.md       | 10 ++--
 docs/cluster_setup.md           | 58 ++++++++++-----------
 docs/faq.md                     | 70 +++++++++++++-------------
 docs/how_to_contribute.md       |  2 +-
 docs/img/FlinkOnYarn.svg        |  4 ++
 docs/img/StratosphereOnYarn.svg |  4 --
 docs/internal_add_operator.md   | 32 ++++++------
 docs/iterations.md              |  2 +-
 docs/java_api_examples.md       | 18 +++----
 docs/java_api_guide.md          | 98 ++++++++++++++++++------------------
 docs/java_api_quickstart.md     | 14 +++---
 docs/local_execution.md         | 12 ++---
 docs/local_setup.md             | 42 ++++++++--------
 docs/run_example_quickstart.md  | 32 ++++++------
 docs/scala_api_examples.md      | 18 +++----
 docs/scala_api_guide.md         | 62 +++++++++++------------
 docs/scala_api_quickstart.md    | 10 ++--
 docs/setup_quickstart.md        | 40 +++++++--------
 docs/spargel_guide.md           |  2 +-
 docs/web_client.md              | 10 ++--
 docs/yarn_setup.md              | 64 +++++++++++------------
 22 files changed, 329 insertions(+), 329 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-flink/blob/fbc93386/docs/cli.md
----------------------------------------------------------------------
diff --git a/docs/cli.md b/docs/cli.md
index 5f6d15b..f716065 100644
--- a/docs/cli.md
+++ b/docs/cli.md
@@ -2,16 +2,16 @@
 title:  "Command-Line Interface"
 ---
 
-Stratosphere provides a command-line interface to run programs that are packaged
+Flink provides a command-line interface to run programs that are packaged
 as JAR files, and control their execution.  The command line interface is part
-of any Stratosphere setup, available in local single node setups and in
-distributed setups. It is located under `<stratosphere-home>/bin/stratosphere`
-and connects by default to the running Stratosphere master (JobManager) that was
+of any Flink setup, available in local single node setups and in
+distributed setups. It is located under `<flink-home>/bin/flink`
+and connects by default to the running Flink master (JobManager) that was
 started from the same installation directory.
 
-A prerequisite to using the command line interface is that the Stratosphere
-master (JobManager) has been started (via `<stratosphere-home>/bin/start-
-local.sh` or `<stratosphere-home>/bin/start-cluster.sh`).
+A prerequisite to using the command line interface is that the Flink
+master (JobManager) has been started (via `<flink-home>/bin/start-
+local.sh` or `<flink-home>/bin/start-cluster.sh`).
 
 The command line can be used to
 
@@ -24,64 +24,64 @@ The command line can be used to
 
 -   Run example program with no arguments.
 
-        ./bin/stratosphere run ./examples/stratosphere-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar
+        ./bin/flink run ./examples/flink-java-examples-{{ site.FLINK_VERSION_STABLE }}-WordCount.jar
 
 -   Run example program with arguments for input and result files
 
-        ./bin/stratosphere run ./examples/stratosphere-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar \
+        ./bin/flink run ./examples/flink-java-examples-{{ site.FLINK_VERSION_STABLE }}-WordCount.jar
\
                                file:///home/user/hamlet.txt file:///home/user/wordcount_out
 
 -   Run example program with parallelism 16 and arguments for input and result files
 
-        ./bin/stratosphere run -p 16 ./examples/stratosphere-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar \
+        ./bin/flink run -p 16 ./examples/flink-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar \
                                 file:///home/user/hamlet.txt file:///home/user/wordcount_out
 
 -   Run example program on a specific JobManager:
 
-        ./bin/stratosphere run -m myJMHost:6123 \
-                               ./examples/stratosphere-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar \
+        ./bin/flink run -m myJMHost:6123 \
+                               ./examples/flink-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar \
                                -file:///home/user/hamlet.txt file:///home/user/wordcount_out
 
 
 -   Display the expected arguments for the WordCount example program:
 
-        ./bin/stratosphere info -d ./examples/stratosphere-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar
+        ./bin/flink info -d ./examples/flink-java-examples-{{ site.FLINK_VERSION_STABLE }}-WordCount.jar
 
 -   Display the optimized execution plan for the WordCount example program as JSON:
 
-        ./bin/stratosphere info -e 
-                                ./examples/stratosphere-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar \
+        ./bin/flink info -e 
+                                ./examples/flink-java-examples-{{ site.FLINK_VERSION_STABLE
}}-WordCount.jar \
                                 file:///home/user/hamlet.txt file:///home/user/wordcount_out
 
 -   List scheduled and running jobs (including their JobIDs):
 
-        ./bin/stratosphere list -s -r
+        ./bin/flink list -s -r
 
 -   Cancel a job:
 
-        ./bin/stratosphere cancel -i <jobID>
+        ./bin/flink cancel -i <jobID>
 
 # Usage
 
 The command line syntax is as follows:
 
 ```
-./stratosphere <ACTION> [OPTIONS] [ARGUMENTS]
+./flink <ACTION> [OPTIONS] [ARGUMENTS]
 
 General options:
      -h,--help      Show the help for the CLI Frontend, or a specific action.
      -v,--verbose   Print more detailed error messages.
 
 
-Action "run" - compiles and submits a Stratosphere program that is given in the form of a
JAR file.
+Action "run" - compiles and submits a Flink program that is given in the form of a JAR file.
 
   "run" options:
 
-     -p,--parallelism <parallelism> The degree of parallelism for the execution. This
value is used unless the program overrides the degree of parallelism on the execution environment
or program plan. If this option is not set, then the execution will use the default parallelism
specified in the stratosphere-conf.yaml file.
+     -p,--parallelism <parallelism> The degree of parallelism for the execution. This
value is used unless the program overrides the degree of parallelism on the execution environment
or program plan. If this option is not set, then the execution will use the default parallelism
specified in the flink-conf.yaml file.
 
      -c,--class <classname>         The class with the entry point (main method, or
getPlan() method). Needs only be specified if the JAR file has no manifest pointing to that
class. See program packaging instructions for details.
 
-     -m,--jobmanager <host:port>    Option to submit the program to a different Stratosphere
master (JobManager).
+     -m,--jobmanager <host:port>    Option to submit the program to a different Flink
master (JobManager).
 
   "run" arguments:
 
@@ -89,7 +89,7 @@ Action "run" - compiles and submits a Stratosphere program that is given
in the
      - All successive arguments are passed to the program's main method (or getPlan() method).
 
 
-Action "info" - displays information about a Stratosphere program.
+Action "info" - displays information about a Flink program.
 
   "info" action arguments:
      -d,--description               Show description of the program, if the main class implements
the 'ProgramDescription' interface.
@@ -100,7 +100,7 @@ Action "info" - displays information about a Stratosphere program.
 
      -c,--class <classname>         The class with the entry point (main method, or
getPlan() method). Needs only be specified if the JAR file has no manifest pointing to that
class. See program packaging instructions for details.
 
-     -m,--jobmanager <host:port>    Option to connect to a different Stratosphere master
(JobManager). Connecting to a master is relevant to compile the execution plan. The option
is only evaluated if used together with the -e option.
+     -m,--jobmanager <host:port>    Option to connect to a different Flink master (JobManager).
Connecting to a master is relevant to compile the execution plan. The option is only evaluated
if used together with the -e option.
 
   "info" arguments:
 
@@ -108,7 +108,7 @@ Action "info" - displays information about a Stratosphere program.
      - All successive arguments are passed to the program's main method (or getPlan() method).
 
 
-Action "list" lists submitted Stratosphere programs.
+Action "list" lists submitted Flink programs.
 
   "list" action arguments:
 
@@ -116,14 +116,14 @@ Action "list" lists submitted Stratosphere programs.
 
      -s,--scheduled                 Show scheduled programs and their JobIDs
 
-     -m,--jobmanager <host:port>    Option to connect to a different Stratosphere master
(JobManager).
+     -m,--jobmanager <host:port>    Option to connect to a different Flink master (JobManager).
 
 
-Action "cancel" cancels a submitted Stratosphere program.
+Action "cancel" cancels a submitted Flink program.
 
   "cancel" action arguments:
 
      -i,--jobid <jobID>             JobID of program to cancel
      
-     -m,--jobmanager <host:port>    Option to connect to a different Stratosphere master
(JobManager).
+     -m,--jobmanager <host:port>    Option to connect to a different Flink master (JobManager).
 ```

http://git-wip-us.apache.org/repos/asf/incubator-flink/blob/fbc93386/docs/cluster_execution.md
----------------------------------------------------------------------
diff --git a/docs/cluster_execution.md b/docs/cluster_execution.md
index 015a656..1d74f7c 100644
--- a/docs/cluster_execution.md
+++ b/docs/cluster_execution.md
@@ -2,7 +2,7 @@
 title:  "Cluster Execution"
 ---
 
-Stratosphere programs can run distributed on clusters of many machines. There
+Flink programs can run distributed on clusters of many machines. There
 are two ways to send a program to a cluster for execution:
 
 # Command Line Interface
@@ -15,14 +15,14 @@ details.
 
 # Remote Environment
 
-The remote environment lets you execute Stratosphere Java programs on a cluster
+The remote environment lets you execute Flink Java programs on a cluster
 directly. The remote environment points to the cluster on which you want to
 execute the program.
 
 ## Maven Dependency
 
 If you are developing your program as a Maven project, you have to add the
-`stratosphere-clients` module using this dependency:
+`flink-clients` module using this dependency:
 
 ```xml
 <dependency>
@@ -62,13 +62,13 @@ takes the path(s) to the JAR file(s).
 # Remote Executor
 
 Similar to the RemoteEnvironment, the RemoteExecutor lets you execute
-Stratosphere programs on a cluster directly. The remote executor accepts a
+Flink programs on a cluster directly. The remote executor accepts a
 *Plan* object, which describes the program as a single executable unit.
 
 ## Maven Dependency
 
 If you are developing your program in a Maven project, you have to add the
-`stratosphere-clients` module using this dependency:
+`flink-clients` module using this dependency:
 
 ```xml
 <dependency>

http://git-wip-us.apache.org/repos/asf/incubator-flink/blob/fbc93386/docs/cluster_setup.md
----------------------------------------------------------------------
diff --git a/docs/cluster_setup.md b/docs/cluster_setup.md
index bc82e96..af97916 100644
--- a/docs/cluster_setup.md
+++ b/docs/cluster_setup.md
@@ -3,10 +3,10 @@ title:  "Cluster Setup"
 ---
 
 This documentation is intended to provide instructions on how to run
-Stratosphere in a fully distributed fashion on a static (but possibly
+Flink in a fully distributed fashion on a static (but possibly
 heterogeneous) cluster.
 
-This involves two steps. First, installing and configuring Stratosphere and
+This involves two steps. First, installing and configuring Flink and
 second installing and configuring the [Hadoop Distributed
 Filesystem](http://hadoop.apache.org/) (HDFS).
 
@@ -14,13 +14,13 @@ Filesystem](http://hadoop.apache.org/) (HDFS).
 
 ## Software Requirements
 
-Stratosphere runs on all *UNIX-like environments*, e.g. **Linux**, **Mac OS X**,
+Flink runs on all *UNIX-like environments*, e.g. **Linux**, **Mac OS X**,
 and **Cygwin** (for Windows) and expects the cluster to consist of **one master
 node** and **one or more worker nodes**. Before you start to setup the system,
 make sure you have the following software installed **on each node**:
 
 - **Java 1.6.x** or higher,
-- **ssh** (sshd must be running to use the Stratosphere scripts that manage
+- **ssh** (sshd must be running to use the Flink scripts that manage
   remote components)
 
 If your cluster does not fulfill these software requirements you will need to
@@ -67,16 +67,16 @@ root       894  0.0  0.0  49260   320 ?        Ss   Jan09   0:13 /usr/sbin/sshd
 In order to start/stop the remote processes, the master node requires access via
 ssh to the worker nodes. It is most convenient to use ssh's public key
 authentication for this. To setup public key authentication, log on to the
-master as the user who will later execute all the Stratosphere components. **The
+master as the user who will later execute all the Flink components. **The
 same user (i.e. a user with the same user name) must also exist on all worker
 nodes**. For the remainder of this instruction we will refer to this user as
-*stratosphere*. Using the super user *root* is highly discouraged for security
+*flink*. Using the super user *root* is highly discouraged for security
 reasons.
 
 Once you logged in to the master node as the desired user, you must generate a
 new public/private key pair. The following command will create a new
 public/private key pair into the *.ssh* directory inside the home directory of
-the user *stratosphere*. See the ssh-keygen man page for more details. Note that
+the user *flink*. See the ssh-keygen man page for more details. Note that
 the private key is not protected by a passphrase.
 
 ```
@@ -110,11 +110,11 @@ worker node from your master node via ssh without a password.
 
 ## Setting JAVA_HOME on each Node
 
-Stratosphere requires the `JAVA_HOME` environment variable to be set on the
+Flink requires the `JAVA_HOME` environment variable to be set on the
 master and all worker nodes and point to the directory of your Java
 installation.
 
-You can set this variable in `conf/stratosphere-conf.yaml` via the
+You can set this variable in `conf/flink-conf.yaml` via the
 `env.java.home` key.
 
 Alternatively, add the following line to your shell profile. If you use the
@@ -136,7 +136,7 @@ echo "PermitUserEnvironment yes" >> /etc/ssh/sshd_config
 
 # Hadoop Distributed Filesystem (HDFS) Setup
 
-The Stratosphere system currently uses the Hadoop Distributed Filesystem (HDFS)
+The Flink system currently uses the Hadoop Distributed Filesystem (HDFS)
 to read and write data in a distributed fashion.
 
 Make sure to have a running HDFS installation. The following instructions are
@@ -148,7 +148,7 @@ many installation guides available online for more detailed instructions.
 
 ## Downloading, Installing, and Configuring HDFS
 
-Similar to the Stratosphere system HDFS runs in a distributed fashion. HDFS
+Similar to the Flink system HDFS runs in a distributed fashion. HDFS
 consists of a **NameNode** which manages the distributed file system's meta
 data. The actual data is stored by one or more **DataNodes**. For the remainder
 of this instruction we assume the HDFS's NameNode component runs on the master
@@ -195,7 +195,7 @@ Guide](http://hadoop.apache.org/docs/r1.2.1/hdfs_user_guide.html) guide.
 Replace *MASTER* with the IP/host name of your master node which runs the
 *NameNode*. *DATAPATH* must be replaced with path to the directory in which the
 actual HDFS data shall be stored on each worker node. Make sure that the
-*stratosphere* user has sufficient permissions to read and write in that
+*flink* user has sufficient permissions to read and write in that
 directory.
 
 After having saved the HDFS configuration file, open the file *conf/slaves* and
@@ -244,31 +244,31 @@ like to point you to the [Hadoop Quick
 Start](http://wiki.apache.org/hadoop/QuickStart)
 guide.
 
-# Stratosphere Setup
+# Flink Setup
 
 Go to the [downloads page](downloads/) and get the ready to run
-package. Make sure to pick the Stratosphere package **matching your Hadoop
+package. Make sure to pick the Flink package **matching your Hadoop
 version**.
 
 After downloading the latest release, copy the archive to your master node and
 extract it:
 
 ```
-tar xzf stratosphere-*.tgz
-cd stratosphere-*
+tar xzf flink-*.tgz
+cd flink-*
 ```
 
 ## Configuring the Cluster
 
-After having extracted the system files, you need to configure Stratosphere for
-the cluster by editing *conf/stratosphere-conf.yaml*.
+After having extracted the system files, you need to configure Flink for
+the cluster by editing *conf/flink-conf.yaml*.
 
 Set the `jobmanager.rpc.address` key to point to your master node. Furthermode
 define the maximum amount of main memory the JVM is allowed to allocate on each
 node by setting the `jobmanager.heap.mb` and `taskmanager.heap.mb` keys.
 
 The value is given in MB. If some worker nodes have more main memory which you
-want to allocate to the Stratosphere system you can overwrite the default value
+want to allocate to the Flink system you can overwrite the default value
 by setting an environment variable `STRATOSPHERE_TM_HEAP` on the respective
 node.
 
@@ -288,9 +288,9 @@ Each entry must be separated by a new line, as in the following example:
 192.168.0.150
 ```
 
-The Stratosphere directory must be available on every worker under the same
+The Flink directory must be available on every worker under the same
 path. Similarly as for HDFS, you can use a shared NSF directory, or copy the
-entire Stratosphere directory to every worker node.
+entire Flink directory to every worker node.
 
 ## Configuring the Network Buffers
 
@@ -328,35 +328,35 @@ parameters:
 
 ## Configuring Temporary I/O Directories
 
-Although Stratosphere aims to process as much data in main memory as possible,
+Although Flink aims to process as much data in main memory as possible,
 it is not uncommon that  more data needs to be processed than memory is
-available. Stratosphere's runtime is designed to  write temporary data to disk
+available. Flink's runtime is designed to  write temporary data to disk
 to handle these situations.
 
 The `taskmanager.tmp.dirs` parameter specifies a list of directories into which
-Stratosphere writes temporary files. The paths of the directories need to be
-separated by ':' (colon character).  Stratosphere will concurrently write (or
+Flink writes temporary files. The paths of the directories need to be
+separated by ':' (colon character).  Flink will concurrently write (or
 read) one temporary file to (from) each configured directory.  This way,
 temporary I/O can be evenly distributed over multiple independent I/O devices
 such as hard disks to improve performance.  To leverage fast I/O devices (e.g.,
 SSD, RAID, NAS), it is possible to specify a directory multiple times.
 
 If the `taskmanager.tmp.dirs` parameter is not explicitly specified,
-Stratosphere writes temporary data to the temporary  directory of the operating
+Flink writes temporary data to the temporary  directory of the operating
 system, such as */tmp* in Linux systems.
 
 Please see the [configuration page](config.html) for details and additional
 configuration options.
 
-## Starting Stratosphere
+## Starting Flink
 
 The following script starts a JobManager on the local node and connects via
 SSH to all worker nodes listed in the *slaves* file to start the
-TaskManager on each node. Now your Stratosphere system is up and
+TaskManager on each node. Now your Flink system is up and
 running. The JobManager running on the local node will now accept jobs
 at the configured RPC port.
 
-Assuming that you are on the master node and inside the Stratosphere directory:
+Assuming that you are on the master node and inside the Flink directory:
 
 ```
 bin/start-cluster.sh

http://git-wip-us.apache.org/repos/asf/incubator-flink/blob/fbc93386/docs/faq.md
----------------------------------------------------------------------
diff --git a/docs/faq.md b/docs/faq.md
index 8600f00..694c156 100644
--- a/docs/faq.md
+++ b/docs/faq.md
@@ -4,25 +4,25 @@ title: "Frequently Asked Questions (FAQ)"
 
 # General
 
-## Is Stratosphere a Hadoop Project?
+## Is Flink a Hadoop Project?
 
-Stratosphere is a data processing system and an alternative to Hadoop's
+Flink is a data processing system and an alternative to Hadoop's
 MapReduce component. It comes with its own runtime, rather than building on top
 of MapReduce. As such, it can work completely independently of the Hadoop
-ecosystem. However, Stratosphere can also access Hadoop's distributed file
+ecosystem. However, Flink can also access Hadoop's distributed file
 system (HDFS) to read and write data, and Hadoop's next-generation resource
-manager (YARN) to provision cluster resources. Since most Stratosphere users are
+manager (YARN) to provision cluster resources. Since most Flink users are
 using Hadoop HDFS to store their data, we ship already the required libraries to
 access HDFS.
 
-## Do I have to install Apache Hadoop to use Stratosphere?
+## Do I have to install Apache Hadoop to use Flink?
 
-No. Stratosphere can run without a Hadoop installation. However, a very common
-setup is to use Stratosphere to analyze data stored in the Hadoop Distributed
+No. Flink can run without a Hadoop installation. However, a very common
+setup is to use Flink to analyze data stored in the Hadoop Distributed
 File System (HDFS). To make these setups work out of the box, we bundle the
-Hadoop client libraries with Stratosphere by default.
+Hadoop client libraries with Flink by default.
 
-Additionally, we provide a special YARN Enabled download of Stratosphere for
+Additionally, we provide a special YARN Enabled download of Flink for
 users with an existing Hadoop YARN cluster. [Apache Hadoop
 YARN](http://hadoop.apache.org/docs/r2.2.0/hadoop-yarn/hadoop-yarn-site/YARN.html) 
 is Hadoop's cluster resource manager that allows to use
@@ -30,13 +30,13 @@ different execution engines next to each other on a cluster.
 
 # Usage
 
-## How do I assess the progress of a Stratosphere program?
+## How do I assess the progress of a Flink program?
 
-There are a multiple of ways to track the progress of a Stratosphere program:
+There are a multiple of ways to track the progress of a Flink program:
 
 - The JobManager (the master of the distributed system) starts a web interface
 to observe program execution. In runs on port 8081 by default (configured in
-`conf/stratosphere-config.yml`).
+`conf/flink-config.yml`).
 - When you start a program from the command line, it will print the status
 changes of all operators as the program progresses through the operations.
 - All status changes are also logged to the JobManager's log file.
@@ -51,10 +51,10 @@ the standard error stream and shown on the console.
 parallel task first failed and caused the other tasks to cancel the execution.
 - Failing tasks and the corresponding exceptions are reported in the log files
 of the master and the worker where the exception occurred
-(`log/stratosphere-<user>-jobmanager-<host>.log` and
-`log/stratosphere-<user>-taskmanager-<host>.log`).
+(`log/flink-<user>-jobmanager-<host>.log` and
+`log/flink-<user>-taskmanager-<host>.log`).
 
-## How do I debug Stratosphere programs?
+## How do I debug Flink programs?
 
 - When you start a program locally with the [LocalExecutor](local_execution.html),
 you can place breakpoints in your functions and debug them like normal
@@ -68,7 +68,7 @@ execution.
 
 ## I get an error message saying that not enough buffers are available. How do I fix this?
 
-If you run Stratosphere in a massively parallel setting (100+ parallel threads),
+If you run Flink in a massively parallel setting (100+ parallel threads),
 you need to adapt the number of network buffers via the config parameter
 `taskmanager.network.numberOfBuffers`.
 As a rule-of-thumb, the number of buffers should be at least
@@ -80,7 +80,7 @@ As a rule-of-thumb, the number of buffers should be at least
 Note: In version _0.4_, the delta iterations limit the solution set to
 records with fixed-length data types. We will  in the next version.
 
-The most common case for these exception is when Stratosphere is set up with the
+The most common case for these exception is when Flink is set up with the
 wrong HDFS version. Because different HDFS versions are often not compatible
 with each other, the connection between the filesystem master and the client
 breaks.
@@ -96,12 +96,12 @@ Call to <host:port> failed on local exception: java.io.EOFException
     at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207)
     at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170)
     at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82)
-    at eu.stratosphere.runtime.fs.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:276
+    at org.apache.flinkruntime.fs.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:276
 ```
 
-Please refer to the [download page](http://stratosphere.eu/downloads/#maven) and
+Please refer to the [download page]({{site.baseurl}}/downloads.html#maven) and
 the {% gh_link README.md master "build instructions" %}
-for details on how to set up Stratosphere for different Hadoop and HDFS versions.
+for details on how to set up Flink for different Hadoop and HDFS versions.
 
 ## My program does not compute the correct result. Why are my custom key types
 are not grouped/joined correctly?
@@ -118,7 +118,7 @@ All data type classes must be public and have a public nullary constructor
 or interfaces. If the classes are internal classes, they must be public and
 static.
 
-## I can't stop Stratosphere with the provided stop-scripts. What can I do?
+## I can't stop Flink with the provided stop-scripts. What can I do?
 
 Stopping the processes sometimes takes a few seconds, because the shutdown may
 do some cleanup work.
@@ -150,7 +150,7 @@ There are two ways to go about this:
 1. See whether you can use less memory inside the functions. For example, use
 arrays of primitive types instead of object types.
 
-2. Reduce the memory that Stratosphere reserves for its own processing. The
+2. Reduce the memory that Flink reserves for its own processing. The
 TaskManager reserves a certain portion of the available memory for sorting,
 hashing, caching, network buffering, etc. That part of the memory is unavailable
 to the user-defined functions. By reserving it, the system can guarantee to not
@@ -178,13 +178,13 @@ output looks like this:
 
 ```
 07:34:27,004 INFO  org.apache.hadoop.yarn.client.api.impl.YarnClientImpl         - Submitted
application application_1395604279745_273123 to ResourceManager at jobtracker-host
-Stratosphere JobManager is now running on worker1:6123
+Flink JobManager is now running on worker1:6123
 JobManager Web Interface: http://jobtracker-host:54311/proxy/application_1295604279745_273123/
-07:34:51,528 INFO  eu.stratosphere.yarn.Client                                   - Application
application_1295604279745_273123 finished with state FINISHED at 1398152089553
-07:34:51,529 INFO  eu.stratosphere.yarn.Client                                   - Killing
the Stratosphere-YARN application.
+07:34:51,528 INFO  org.apache.flinkyarn.Client                                   - Application
application_1295604279745_273123 finished with state FINISHED at 1398152089553
+07:34:51,529 INFO  org.apache.flinkyarn.Client                                   - Killing
the Flink-YARN application.
 07:34:51,529 INFO  org.apache.hadoop.yarn.client.api.impl.YarnClientImpl         - Killing
application application_1295604279745_273123
-07:34:51,534 INFO  eu.stratosphere.yarn.Client                                   - Deleting
files in hdfs://user/marcus/.stratosphere/application_1295604279745_273123
-07:34:51,559 INFO  eu.stratosphere.yarn.Client                                   - YARN Client
is shutting down
+07:34:51,534 INFO  org.apache.flinkyarn.Client                                   - Deleting
files in hdfs://user/marcus/.flink/application_1295604279745_273123
+07:34:51,559 INFO  org.apache.flinkyarn.Client                                   - YARN Client
is shutting down
 ```
 
 The problem here is that the Application Master (AM) is stopping and the YARN client assumes
that the application has finished.
@@ -255,30 +255,30 @@ Exception in thread "main" org.apache.hadoop.security.AccessControlException:
Pe
   at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:2021)
   at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1989)
   at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1954)
-  at eu.stratosphere.yarn.Utils.setupLocalResource(Utils.java:176)
-  at eu.stratosphere.yarn.Client.run(Client.java:362)
-  at eu.stratosphere.yarn.Client.main(Client.java:568)
+  at org.apache.flinkyarn.Utils.setupLocalResource(Utils.java:176)
+  at org.apache.flinkyarn.Client.run(Client.java:362)
+  at org.apache.flinkyarn.Client.main(Client.java:568)
 ```
 
 The reason for this error is, that the home directory of the user **in HDFS**
 has the wrong permissions. The user (in this case `robert`) can not create
 directories in his own home directory.
 
-Stratosphere creates a `.stratosphere/` directory in the users home directory
-where it stores the Stratosphere jar and configuration file.
+Flink creates a `.flink/` directory in the users home directory
+where it stores the Flink jar and configuration file.
 
 # Features
 
-## What kind of fault-tolerance does Stratosphere provide?
+## What kind of fault-tolerance does Flink provide?
 
 Stratospere can restart failed jobs. Mid-query fault tolerance will go into the
 open source project in the next versions.
 
 ## Are Hadoop-like utilities, such as Counters and the DistributedCache supported?
 
-[Stratosphere's Accumulators](java_api_guide.html#accumulators-&-counters) work very
similar like
+[Flink's Accumulators](java_api_guide.html#accumulators-&-counters) work very similar
like
 [Hadoop's counters, but are more powerful.
 
-Stratosphere has a {% gh_link /stratosphere-core/src/main/java/eu/stratosphere/api/common/cache/DistributedCache.java
"Distributed Cache" %} that is deeply integrated with the APIs. Please refer to the {% gh_link
/stratosphere-java/src/main/java/eu/stratosphere/api/java/ExecutionEnvironment.java#L561 "JavaDocs"
%} for details on how to use it.
+Flink has a {% gh_link /flink-core/src/main/java/org/apache/flink/api/common/cache/DistributedCache.java
"Distributed Cache" %} that is deeply integrated with the APIs. Please refer to the {% gh_link
/flink-java/src/main/java/org/apache/flink/api/java/ExecutionEnvironment.java#L561 "JavaDocs"
%} for details on how to use it.
 
 In order to make data sets available on all tasks, we encourage you to use [Broadcast Variables](java_api_guide.html#broadcast_variables)
instead. They are more efficient and easier to use than the distributed cache.

http://git-wip-us.apache.org/repos/asf/incubator-flink/blob/fbc93386/docs/how_to_contribute.md
----------------------------------------------------------------------
diff --git a/docs/how_to_contribute.md b/docs/how_to_contribute.md
index ef23c38..4b532d9 100644
--- a/docs/how_to_contribute.md
+++ b/docs/how_to_contribute.md
@@ -37,7 +37,7 @@ git clone https://github.com/<your-user-name>/incubator-flink.git
 git checkout -b myBranch master
 ```
 
-4. Now you can create your changes, compile the code, and validate the changes. Here are
some pointers on how to [set up the Eclipse IDE for development](https://github.com/apache/incubator-flink/#eclipse-setup-and-debugging),
and how to [build the code](https://github.com/apache/incubator-flink/#build-stratosphere).
+4. Now you can create your changes, compile the code, and validate the changes. Here are
some pointers on how to [set up the Eclipse IDE for development](https://github.com/apache/incubator-flink/#eclipse-setup-and-debugging),
and how to [build the code](https://github.com/apache/incubator-flink/#build-flink).
 
 5. After you have finalized your contribution, verify the compliance with the contribution
guidelines (see below), and commit them. To make the changes easily mergeable, please rebase
them to the latest version of the main repositories master branch. Assuming you created a
topic branch (step 3), you can follow this sequence of commands to do that:
 Switch to the master branch, update it to the latest revision, switch back to your topic
branch, and rebase it on top of the master branch.


Mime
View raw message