From bdbc99f792f673fafc82ac3bda3dd2bb5d497eec Mon Sep 17 00:00:00 2001 From: sirko bretschneider Date: Sun, 1 Oct 2017 19:44:20 +0200 Subject: [PATCH] [FLINK-7744][docs] Add missing top links to documentation --- docs/dev/batch/connectors.md | 2 ++ docs/dev/batch/dataset_transformations.md | 2 ++ docs/dev/batch/examples.md | 2 ++ docs/dev/batch/hadoop_compatibility.md | 2 ++ docs/dev/batch/iterations.md | 2 ++ docs/dev/batch/zip_elements_guide.md | 2 ++ docs/dev/best_practices.md | 2 +- docs/dev/cluster_execution.md | 2 ++ docs/dev/connectors/cassandra.md | 2 ++ docs/dev/connectors/elasticsearch.md | 2 ++ docs/dev/connectors/filesystem_sink.md | 2 ++ docs/dev/connectors/kafka.md | 2 ++ docs/dev/connectors/kinesis.md | 2 ++ docs/dev/connectors/nifi.md | 2 ++ docs/dev/connectors/rabbitmq.md | 2 ++ docs/dev/connectors/twitter.md | 2 ++ docs/dev/custom_serializers.md | 2 ++ docs/dev/event_time.md | 2 ++ docs/dev/event_timestamp_extractors.md | 2 ++ docs/dev/event_timestamps_watermarks.md | 2 +- docs/dev/java8.md | 2 ++ docs/dev/libs/cep.md | 2 ++ docs/dev/libs/gelly/bipartite_graph.md | 2 ++ docs/dev/libs/ml/als.md | 2 ++ docs/dev/libs/ml/contribution_guide.md | 2 ++ docs/dev/libs/ml/cross_validation.md | 2 ++ docs/dev/libs/ml/distance_metrics.md | 2 ++ docs/dev/libs/ml/knn.md | 2 ++ docs/dev/libs/ml/min_max_scaler.md | 2 ++ docs/dev/libs/ml/multiple_linear_regression.md | 2 ++ docs/dev/libs/ml/optimization.md | 2 ++ docs/dev/libs/ml/pipelines.md | 4 +++- docs/dev/libs/ml/polynomial_features.md | 2 ++ docs/dev/libs/ml/quickstart.md | 2 ++ docs/dev/libs/ml/sos.md | 2 ++ docs/dev/libs/ml/standard_scaler.md | 2 ++ docs/dev/libs/ml/svm.md | 2 ++ docs/dev/libs/storm_compatibility.md | 2 ++ docs/dev/linking.md | 2 ++ docs/dev/local_execution.md | 2 ++ docs/dev/migration.md | 2 ++ docs/dev/scala_api_extensions.md | 2 ++ docs/dev/scala_shell.md | 2 ++ docs/dev/stream/operators/asyncio.md | 1 + docs/dev/stream/operators/windows.md | 2 ++ docs/dev/stream/side_output.md | 2 ++ docs/dev/stream/state/custom_serialization.md | 2 ++ docs/dev/stream/state/queryable_state.md | 2 ++ docs/dev/stream/state/state.md | 4 +++- docs/dev/stream/state/state_backends.md | 2 ++ docs/dev/stream/testing.md | 2 ++ docs/dev/types_serialization.md | 2 ++ docs/internals/components.md | 2 ++ docs/internals/filesystems.md | 1 + docs/internals/ide_setup.md | 2 ++ docs/internals/job_scheduling.md | 2 ++ docs/internals/stream_checkpointing.md | 2 ++ docs/internals/task_lifecycle.md | 2 ++ docs/monitoring/application_profiling.md | 2 ++ docs/monitoring/back_pressure.md | 2 ++ docs/monitoring/checkpoint_monitoring.md | 2 ++ docs/monitoring/debugging_classloading.md | 1 + docs/monitoring/debugging_event_time.md | 1 + docs/monitoring/historyserver.md | 2 ++ docs/monitoring/logging.md | 2 ++ docs/monitoring/rest_api.md | 2 ++ docs/ops/cli.md | 2 ++ docs/ops/config.md | 2 ++ docs/ops/deployment/aws.md | 2 ++ docs/ops/deployment/gce_setup.md | 2 ++ docs/ops/deployment/mapr_setup.md | 2 ++ docs/ops/deployment/mesos.md | 2 ++ docs/ops/deployment/yarn_setup.md | 2 ++ docs/ops/jobmanager_high_availability.md | 2 ++ docs/ops/production_ready.md | 2 ++ docs/ops/security-kerberos.md | 2 ++ docs/ops/security-ssl.md | 1 + docs/ops/state/checkpoints.md | 2 ++ docs/ops/state/large_state_tuning.md | 2 ++ docs/ops/state/savepoints.md | 2 ++ docs/ops/state/state_backends.md | 2 ++ docs/ops/upgrading.md | 2 ++ docs/quickstart/java_api_quickstart.md | 2 ++ docs/quickstart/run_example_quickstart.md | 2 ++ docs/quickstart/scala_api_quickstart.md | 2 ++ docs/quickstart/setup_quickstart.md | 2 ++ docs/search-results.md | 2 ++ 87 files changed, 169 insertions(+), 4 deletions(-) diff --git a/docs/dev/batch/connectors.md b/docs/dev/batch/connectors.md index 93bbf72c2706f..388b59925f3c1 100644 --- a/docs/dev/batch/connectors.md +++ b/docs/dev/batch/connectors.md @@ -224,3 +224,5 @@ The example shows how to access an Azure table and turn data into Flink's `DataS ## Access MongoDB This [GitHub repository documents how to use MongoDB with Apache Flink (starting from 0.7-incubating)](https://github.com/okkam-it/flink-mongodb-test). + +{% top %} diff --git a/docs/dev/batch/dataset_transformations.md b/docs/dev/batch/dataset_transformations.md index c322f22780a86..d63ee88ccd1f9 100644 --- a/docs/dev/batch/dataset_transformations.md +++ b/docs/dev/batch/dataset_transformations.md @@ -2338,3 +2338,5 @@ Not supported. + +{% top %} diff --git a/docs/dev/batch/examples.md b/docs/dev/batch/examples.md index fe478f8d1eba5..508beef0fa31e 100644 --- a/docs/dev/batch/examples.md +++ b/docs/dev/batch/examples.md @@ -517,3 +517,5 @@ CC = gcc ~~~bash ./dbgen -T o -s 1 ~~~ + +{% top %} diff --git a/docs/dev/batch/hadoop_compatibility.md b/docs/dev/batch/hadoop_compatibility.md index 9548c292b142b..bbeea09e370dc 100644 --- a/docs/dev/batch/hadoop_compatibility.md +++ b/docs/dev/batch/hadoop_compatibility.md @@ -246,3 +246,5 @@ result.output(hadoopOF); // Execute Program env.execute("Hadoop WordCount"); ~~~ + +{% top %} diff --git a/docs/dev/batch/iterations.md b/docs/dev/batch/iterations.md index 73a1d576adce9..67f2615cd0c65 100644 --- a/docs/dev/batch/iterations.md +++ b/docs/dev/batch/iterations.md @@ -208,3 +208,5 @@ We referred to each execution of the step function of an iteration operator as *

Supersteps

+ +{% top %} diff --git a/docs/dev/batch/zip_elements_guide.md b/docs/dev/batch/zip_elements_guide.md index a5c65c55cbecd..0ec5f84fada41 100644 --- a/docs/dev/batch/zip_elements_guide.md +++ b/docs/dev/batch/zip_elements_guide.md @@ -124,3 +124,5 @@ env.execute() may yield the tuples: (0,G), (1,A), (2,H), (3,B), (5,C), (7,D), (9,E), (11,F) [Back to top](#top) + +{% top %} diff --git a/docs/dev/best_practices.md b/docs/dev/best_practices.md index d01ea0f9fdd74..2a1d32e9ef314 100644 --- a/docs/dev/best_practices.md +++ b/docs/dev/best_practices.md @@ -317,4 +317,4 @@ Note that you need to explicitly set the `lib/` directory when using a per-job Y The command to submit Flink on YARN with a custom logger is: `./bin/flink run -yt $FLINK_HOME/lib <... remaining arguments ...>` - +{% top %} diff --git a/docs/dev/cluster_execution.md b/docs/dev/cluster_execution.md index 03af63744ad46..f1d84e1b67b1f 100644 --- a/docs/dev/cluster_execution.md +++ b/docs/dev/cluster_execution.md @@ -81,3 +81,5 @@ public static void main(String[] args) throws Exception { Note that the program contains custom user code and hence requires a JAR file with the classes of the code attached. The constructor of the remote environment takes the path(s) to the JAR file(s). + +{% top %} diff --git a/docs/dev/connectors/cassandra.md b/docs/dev/connectors/cassandra.md index c897779f29e73..12b7ce7bd984e 100644 --- a/docs/dev/connectors/cassandra.md +++ b/docs/dev/connectors/cassandra.md @@ -152,3 +152,5 @@ public class Pojo implements Serializable { {% endhighlight %} + +{% top %} diff --git a/docs/dev/connectors/elasticsearch.md b/docs/dev/connectors/elasticsearch.md index 3fba7f01fea8b..b6ee63c0449f8 100644 --- a/docs/dev/connectors/elasticsearch.md +++ b/docs/dev/connectors/elasticsearch.md @@ -471,3 +471,5 @@ adding the following to the Maven POM file in the plugins section: ~~~ + +{% top %} diff --git a/docs/dev/connectors/filesystem_sink.md b/docs/dev/connectors/filesystem_sink.md index 2d48876ba107b..4e1f68a9f3abc 100644 --- a/docs/dev/connectors/filesystem_sink.md +++ b/docs/dev/connectors/filesystem_sink.md @@ -135,3 +135,5 @@ because of the batch size. For in-depth information, please refer to the JavaDoc for [BucketingSink](http://flink.apache.org/docs/latest/api/java/org/apache/flink/streaming/connectors/fs/bucketing/BucketingSink.html). + +{% top %} diff --git a/docs/dev/connectors/kafka.md b/docs/dev/connectors/kafka.md index f95c8c09ee477..b8e038d867209 100644 --- a/docs/dev/connectors/kafka.md +++ b/docs/dev/connectors/kafka.md @@ -594,3 +594,5 @@ When using standalone Flink deployment, you can also use `SASL_SSL`; please see For more information on Flink configuration for Kerberos security, please see [here]({{ site.baseurl}}/ops/config.html). You can also find [here]({{ site.baseurl}}/ops/security-kerberos.html) further details on how Flink internally setups Kerberos-based security. + +{% top %} diff --git a/docs/dev/connectors/kinesis.md b/docs/dev/connectors/kinesis.md index aa20d3f8be3e4..2c8b88a6797a7 100644 --- a/docs/dev/connectors/kinesis.md +++ b/docs/dev/connectors/kinesis.md @@ -367,3 +367,5 @@ producerConfig.put(AWSConfigConstants.AWS_ENDPOINT, "http://localhost:4567"); {% endhighlight %} + +{% top %} diff --git a/docs/dev/connectors/nifi.md b/docs/dev/connectors/nifi.md index dbc1e8a915dba..392b173731c0c 100644 --- a/docs/dev/connectors/nifi.md +++ b/docs/dev/connectors/nifi.md @@ -136,3 +136,5 @@ streamExecEnv.addSink(nifiSink) More information about [Apache NiFi](https://nifi.apache.org) Site-to-Site Protocol can be found [here](https://nifi.apache.org/docs/nifi-docs/html/user-guide.html#site-to-site) + +{% top %} diff --git a/docs/dev/connectors/rabbitmq.md b/docs/dev/connectors/rabbitmq.md index 578089281b09f..c3ad4b789281a 100644 --- a/docs/dev/connectors/rabbitmq.md +++ b/docs/dev/connectors/rabbitmq.md @@ -171,3 +171,5 @@ stream.addSink(new RMQSink[String]( More about RabbitMQ can be found [here](http://www.rabbitmq.com/). + +{% top %} diff --git a/docs/dev/connectors/twitter.md b/docs/dev/connectors/twitter.md index 0cded6a51dc73..a563be6e726f3 100644 --- a/docs/dev/connectors/twitter.md +++ b/docs/dev/connectors/twitter.md @@ -83,3 +83,5 @@ The `TwitterExample` class in the `flink-examples-streaming` package shows a ful By default, the `TwitterSource` uses the `StatusesSampleEndpoint`. This endpoint returns a random sample of Tweets. There is a `TwitterSource.EndpointInitializer` interface allowing users to provide a custom endpoint. + +{% top %} diff --git a/docs/dev/custom_serializers.md b/docs/dev/custom_serializers.md index ddfc2ee0e5f1e..2c12e6ea9eaea 100644 --- a/docs/dev/custom_serializers.md +++ b/docs/dev/custom_serializers.md @@ -122,3 +122,5 @@ that makes sure the user code classloader is used. Please refer to [FLINK-6025](https://issues.apache.org/jira/browse/FLINK-6025) for more details. + +{% top %} diff --git a/docs/dev/event_time.md b/docs/dev/event_time.md index 70a78122b9366..a3e697d8f7e6b 100644 --- a/docs/dev/event_time.md +++ b/docs/dev/event_time.md @@ -213,3 +213,5 @@ with late elements in event time windows. Please refer to the [Debugging Windows & Event Time]({{ site.baseurl }}/monitoring/debugging_event_time.html) section for debugging watermarks at runtime. + +{% top %} diff --git a/docs/dev/event_timestamp_extractors.md b/docs/dev/event_timestamp_extractors.md index b2704919cf3dc..01b3634867edf 100644 --- a/docs/dev/event_timestamp_extractors.md +++ b/docs/dev/event_timestamp_extractors.md @@ -105,3 +105,5 @@ val withTimestampsAndWatermarks = stream.assignTimestampsAndWatermarks(new Bound {% endhighlight %} + +{% top %} diff --git a/docs/dev/event_timestamps_watermarks.md b/docs/dev/event_timestamps_watermarks.md index f58f7055cb76d..961948fa0062f 100644 --- a/docs/dev/event_timestamps_watermarks.md +++ b/docs/dev/event_timestamps_watermarks.md @@ -371,4 +371,4 @@ val stream: DataStream[MyType] = env.addSource(kafkaSource) Generating Watermarks with awareness for Kafka-partitions - +{% top %} diff --git a/docs/dev/java8.md b/docs/dev/java8.md index 1912fb1e1c708..df1e0885d7b42 100644 --- a/docs/dev/java8.md +++ b/docs/dev/java8.md @@ -194,3 +194,5 @@ final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment(); env.fromElements(1, 2, 3).map((in) -> new Tuple1(" " + in)).print(); env.execute(); ~~~ + +{% top %} diff --git a/docs/dev/libs/cep.md b/docs/dev/libs/cep.md index 8b16ac40e5f04..0373155524c53 100644 --- a/docs/dev/libs/cep.md +++ b/docs/dev/libs/cep.md @@ -1639,3 +1639,5 @@ the looping patterns, multiple input events can match a single (looping) pattern 3. The `followedBy()` in Flink 1.1 and 1.2 implied `non-deterministic relaxed contiguity` (see [here](#conditions-on-contiguity)). In Flink 1.3 this has changed and `followedBy()` implies `relaxed contiguity`, while `followedByAny()` should be used if `non-deterministic relaxed contiguity` is required. + +{% top %} diff --git a/docs/dev/libs/gelly/bipartite_graph.md b/docs/dev/libs/gelly/bipartite_graph.md index ac57e3bc94598..3aec8ecb61d16 100644 --- a/docs/dev/libs/gelly/bipartite_graph.md +++ b/docs/dev/libs/gelly/bipartite_graph.md @@ -183,3 +183,5 @@ Graph> graph bipartiteG {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/als.md b/docs/dev/libs/ml/als.md index a0ef78accaf05..87c80f87ff7c9 100644 --- a/docs/dev/libs/ml/als.md +++ b/docs/dev/libs/ml/als.md @@ -173,3 +173,5 @@ val testingDS: DataSet[(Int, Int)] = env.readCsvFile[(Int, Int)](pathToData) // Calculate the ratings according to the matrix factorization val predictedRatings = als.predict(testingDS) {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/contribution_guide.md b/docs/dev/libs/ml/contribution_guide.md index 992232fada1aa..b30c53ecac388 100644 --- a/docs/dev/libs/ml/contribution_guide.md +++ b/docs/dev/libs/ml/contribution_guide.md @@ -104,3 +104,5 @@ See `docs/_include/latex_commands.html` for the complete list of predefined late Once you have implemented the algorithm with adequate test coverage and added documentation, you are ready to open a pull request. Details of how to open a pull request can be found [here](http://flink.apache.org/how-to-contribute.html#contributing-code--documentation). + +{% top %} diff --git a/docs/dev/libs/ml/cross_validation.md b/docs/dev/libs/ml/cross_validation.md index 943c492d2bf86..ef3d2ffc177c1 100644 --- a/docs/dev/libs/ml/cross_validation.md +++ b/docs/dev/libs/ml/cross_validation.md @@ -169,3 +169,5 @@ val dataKFolded: Array[TrainTestDataSet] = Splitter.kFoldSplit(data, 10) // create an array of 5 datasets val dataMultiRandom: Array[DataSet[T]] = Splitter.multiRandomSplit(data, Array(0.5, 0.1, 0.1, 0.1, 0.1)) {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/distance_metrics.md b/docs/dev/libs/ml/distance_metrics.md index 1dbd0024bb9b1..3119479080673 100644 --- a/docs/dev/libs/ml/distance_metrics.md +++ b/docs/dev/libs/ml/distance_metrics.md @@ -105,3 +105,5 @@ object MyDistance { val myMetric = MyDistance() {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/knn.md b/docs/dev/libs/ml/knn.md index 0d3ca9ad32708..43f8d13779aa0 100644 --- a/docs/dev/libs/ml/knn.md +++ b/docs/dev/libs/ml/knn.md @@ -142,3 +142,5 @@ val result = knn.predict(testingSet).collect() {% endhighlight %} For more details on the computing KNN with and without and quadtree, here is a presentation: [http://danielblazevski.github.io/](http://danielblazevski.github.io/) + +{% top %} diff --git a/docs/dev/libs/ml/min_max_scaler.md b/docs/dev/libs/ml/min_max_scaler.md index 35376c3771a02..c44a875abcfd3 100644 --- a/docs/dev/libs/ml/min_max_scaler.md +++ b/docs/dev/libs/ml/min_max_scaler.md @@ -110,3 +110,5 @@ minMaxscaler.fit(dataSet) // Scale the provided data set to have min=-1.0 and max=1.0 val scaledDS = minMaxscaler.transform(dataSet) {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/multiple_linear_regression.md b/docs/dev/libs/ml/multiple_linear_regression.md index a5737eb4afe62..c6b7ed65d66fb 100644 --- a/docs/dev/libs/ml/multiple_linear_regression.md +++ b/docs/dev/libs/ml/multiple_linear_regression.md @@ -150,3 +150,5 @@ mlr.fit(trainingDS) // Calculate the predictions for the test data val predictions = mlr.predict(testingDS) {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/optimization.md b/docs/dev/libs/ml/optimization.md index 739d912759a2a..1e3bd2ae1db4e 100644 --- a/docs/dev/libs/ml/optimization.md +++ b/docs/dev/libs/ml/optimization.md @@ -417,3 +417,5 @@ val trainingDS: DataSet[LabeledVector] = ... // Optimize the weights, according to the provided data val weightDS = sgd.optimize(trainingDS) {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/pipelines.md b/docs/dev/libs/ml/pipelines.md index e0f7d828f241e..514d557f4f4e0 100644 --- a/docs/dev/libs/ml/pipelines.md +++ b/docs/dev/libs/ml/pipelines.md @@ -438,4 +438,6 @@ object MeanTransformer { {% endhighlight %} If we wanted to implement a `Predictor` instead of a `Transformer`, then we would have to provide a `FitOperation`, too. -Moreover, a `Predictor` requires a `PredictOperation` which implements how predictions are calculated from testing data. +Moreover, a `Predictor` requires a `PredictOperation` which implements how predictions are calculated from testing data. + +{% top %} diff --git a/docs/dev/libs/ml/polynomial_features.md b/docs/dev/libs/ml/polynomial_features.md index 676c132362f62..5654ec7483f45 100644 --- a/docs/dev/libs/ml/polynomial_features.md +++ b/docs/dev/libs/ml/polynomial_features.md @@ -106,3 +106,5 @@ val pipeline = polyFeatures.chainPredictor(mlr) // train the model pipeline.fit(trainingDS) {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/quickstart.md b/docs/dev/libs/ml/quickstart.md index 5dff6bb015720..ea6f80497553d 100644 --- a/docs/dev/libs/ml/quickstart.md +++ b/docs/dev/libs/ml/quickstart.md @@ -241,3 +241,5 @@ coordinate ascent.* Advances in Neural Information Processing Systems. 2014. [3] Hsu, Chih-Wei, Chih-Chung Chang, and Chih-Jen Lin. *A practical guide to support vector classification.* 2003. + +{% top %} diff --git a/docs/dev/libs/ml/sos.md b/docs/dev/libs/ml/sos.md index 22f4c3049b912..6f117e0a28e6f 100644 --- a/docs/dev/libs/ml/sos.md +++ b/docs/dev/libs/ml/sos.md @@ -118,3 +118,5 @@ outputVector.foreach(output => expectedOutputVector(output._1) should be(output. [1]J.H.M. Janssens, F. Huszar, E.O. Postma, and H.J. van den Herik. *Stochastic Outlier Selection*. Technical Report TiCC TR 2012-001, Tilburg University, Tilburg, the Netherlands, 2012. + +{% top %} diff --git a/docs/dev/libs/ml/standard_scaler.md b/docs/dev/libs/ml/standard_scaler.md index 5104d3c62cee0..cdfc6a059b099 100644 --- a/docs/dev/libs/ml/standard_scaler.md +++ b/docs/dev/libs/ml/standard_scaler.md @@ -111,3 +111,5 @@ scaler.fit(dataSet) // Scale the provided data set to have mean=10.0 and std=2.0 val scaledDS = scaler.transform(dataSet) {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/ml/svm.md b/docs/dev/libs/ml/svm.md index 34fa1ec6c42c3..2fa9e0a9309c6 100644 --- a/docs/dev/libs/ml/svm.md +++ b/docs/dev/libs/ml/svm.md @@ -218,3 +218,5 @@ val testingDS: DataSet[Vector] = env.readLibSVM(pathToTestingFile).map(_.vector) val predictionDS: DataSet[(Vector, Double)] = svm.predict(testingDS) {% endhighlight %} + +{% top %} diff --git a/docs/dev/libs/storm_compatibility.md b/docs/dev/libs/storm_compatibility.md index 6b24dc00bfa0a..4f499f10292ec 100644 --- a/docs/dev/libs/storm_compatibility.md +++ b/docs/dev/libs/storm_compatibility.md @@ -285,3 +285,5 @@ Compare `pom.xml` to see how both jars are built. Furthermore, there is one example for whole Storm topologies (`WordCount-StormTopology.jar`). You can run each of those examples via `bin/flink run .jar`. The correct entry point class is contained in each jar's manifest file. + +{% top %} diff --git a/docs/dev/linking.md b/docs/dev/linking.md index 05926178e6dad..78ef54494831f 100644 --- a/docs/dev/linking.md +++ b/docs/dev/linking.md @@ -92,3 +92,5 @@ the following to your plugins section. ~~~ Now when running `mvn clean package` the produced jar includes the required dependencies. + +{% top %} diff --git a/docs/dev/local_execution.md b/docs/dev/local_execution.md index cf89956159695..326d5157fe3a7 100644 --- a/docs/dev/local_execution.md +++ b/docs/dev/local_execution.md @@ -123,3 +123,5 @@ public static void main(String[] args) throws Exception { The `flink-examples-batch` module contains a full example, called `CollectionExecutionExample`. Please note that the execution of the collection-based Flink programs is only possible on small data, which fits into the JVM heap. The execution on collections is not multi-threaded, only one thread is used. + +{% top %} diff --git a/docs/dev/migration.md b/docs/dev/migration.md index 3369a2c01b87b..5ac69617a8490 100644 --- a/docs/dev/migration.md +++ b/docs/dev/migration.md @@ -476,3 +476,5 @@ val window2 = source {% endhighlight %} + +{% top %} diff --git a/docs/dev/scala_api_extensions.md b/docs/dev/scala_api_extensions.md index 283f50bf48222..41836f9c09959 100644 --- a/docs/dev/scala_api_extensions.md +++ b/docs/dev/scala_api_extensions.md @@ -406,3 +406,5 @@ object Main { } } {% endhighlight %} + +{% top %} diff --git a/docs/dev/scala_shell.md b/docs/dev/scala_shell.md index a8d1b748efc04..bfd3133f6cf21 100644 --- a/docs/dev/scala_shell.md +++ b/docs/dev/scala_shell.md @@ -191,3 +191,5 @@ Starts Flink scala shell connecting to a yarn cluster -h | --help Prints this usage text ~~~ + +{% top %} diff --git a/docs/dev/stream/operators/asyncio.md b/docs/dev/stream/operators/asyncio.md index c5bafa16cdc9e..32945e4ac868c 100644 --- a/docs/dev/stream/operators/asyncio.md +++ b/docs/dev/stream/operators/asyncio.md @@ -251,3 +251,4 @@ For example, the following patterns result in a blocking `asyncInvoke(...)` func - Blocking/waiting on the future-type objects returned by an aynchronous client inside the `asyncInvoke(...)` method +{% top %} diff --git a/docs/dev/stream/operators/windows.md b/docs/dev/stream/operators/windows.md index b825876e734e4..0ecae0cb260f9 100644 --- a/docs/dev/stream/operators/windows.md +++ b/docs/dev/stream/operators/windows.md @@ -1165,3 +1165,5 @@ Windows can be defined over long periods of time (such as days, weeks, or months 2. `FoldFunction` and `ReduceFunction` can significantly reduce the storage requirements, as they eagerly aggregate elements and store only one value per window. In contrast, just using a `WindowFunction` requires accumulating all elements. 3. Using an `Evictor` prevents any pre-aggregation, as all the elements of a window have to be passed through the evictor before applying the computation (see [Evictors](#evictors)). + +{% top %} diff --git a/docs/dev/stream/side_output.md b/docs/dev/stream/side_output.md index da76af407a860..0f144d4196aa3 100644 --- a/docs/dev/stream/side_output.md +++ b/docs/dev/stream/side_output.md @@ -136,3 +136,5 @@ val sideOutputStream: DataStream[String] = mainDataStream.getSideOutput(outputTa {% endhighlight %} + +{% top %} diff --git a/docs/dev/stream/state/custom_serialization.md b/docs/dev/stream/state/custom_serialization.md index fbb7b83a8a791..ca6b07dd00ae1 100644 --- a/docs/dev/stream/state/custom_serialization.md +++ b/docs/dev/stream/state/custom_serialization.md @@ -186,3 +186,5 @@ fundamental components to compatibility checks on upgraded serializers and would is not present. Since configuration snapshots are written to checkpoints using custom serialization, the implementation of the class is free to be changed, as long as compatibility of the configuration change is handled using the versioning mechanisms in `TypeSerializerConfigSnapshot`. + +{% top %} diff --git a/docs/dev/stream/state/queryable_state.md b/docs/dev/stream/state/queryable_state.md index bd0d7fbef4851..4bbc043de305c 100644 --- a/docs/dev/stream/state/queryable_state.md +++ b/docs/dev/stream/state/queryable_state.md @@ -290,3 +290,5 @@ more robust with asks and acknowledgements. * The server and client keep track of statistics for queries. These are currently disabled by default as they would not be exposed anywhere. As soon as there is better support to publish these numbers via the Metrics system, we should enable the stats. + +{% top %} diff --git a/docs/dev/stream/state/state.md b/docs/dev/stream/state/state.md index f280ceb651e9b..0f80a9dff6a43 100644 --- a/docs/dev/stream/state/state.md +++ b/docs/dev/stream/state/state.md @@ -593,4 +593,6 @@ class CounterSource -Some operators might need the information when a checkpoint is fully acknowledged by Flink to communicate that with the outside world. In this case see the `org.apache.flink.runtime.state.CheckpointListener` interface. \ No newline at end of file +Some operators might need the information when a checkpoint is fully acknowledged by Flink to communicate that with the outside world. In this case see the `org.apache.flink.runtime.state.CheckpointListener` interface. + +{% top %} diff --git a/docs/dev/stream/state/state_backends.md b/docs/dev/stream/state/state_backends.md index 1357f2e5a2305..8e32f8ea453a5 100644 --- a/docs/dev/stream/state/state_backends.md +++ b/docs/dev/stream/state/state_backends.md @@ -44,3 +44,5 @@ env.setStateBackend(...) {% endhighlight %} + +{% top %} diff --git a/docs/dev/stream/testing.md b/docs/dev/stream/testing.md index 44f5cfd5d9a12..e5bc02414244c 100644 --- a/docs/dev/stream/testing.md +++ b/docs/dev/stream/testing.md @@ -261,3 +261,5 @@ Another approach is to write a unit test using the Flink internal testing utilit For an example of how to do that please have a look at the `org.apache.flink.streaming.runtime.operators.windowing.WindowOperatorTest` also in the `flink-streaming-java` module. Be aware that `AbstractStreamOperatorTestHarness` is currently not a part of public API and can be subject to change. + +{% top %} diff --git a/docs/dev/types_serialization.md b/docs/dev/types_serialization.md index 0d68a51b59831..00d03635b4057 100644 --- a/docs/dev/types_serialization.md +++ b/docs/dev/types_serialization.md @@ -360,3 +360,5 @@ The parameters provide additional information about the type itself as well as t If your type contains generic parameters that might need to be derived from the input type of a Flink function, make sure to also implement `org.apache.flink.api.common.typeinfo.TypeInformation#getGenericParameters` for a bidirectional mapping of generic parameters to type information. + +{% top %} diff --git a/docs/internals/components.md b/docs/internals/components.md index cf3b6597135ec..e85183b92a1c6 100644 --- a/docs/internals/components.md +++ b/docs/internals/components.md @@ -57,3 +57,5 @@ You can click on the components in the figure to learn more. + +{% top %} diff --git a/docs/internals/filesystems.md b/docs/internals/filesystems.md index 427251ae355f8..5ffd766b86235 100644 --- a/docs/internals/filesystems.md +++ b/docs/internals/filesystems.md @@ -136,3 +136,4 @@ The `FSDataOutputStream` and `FSDataOutputStream` implementations are strictly * Instances of the streams should also not be passed between threads in between read or write operations, because there are no guarantees about the visibility of operations across threads (many operations do not create memory fences). +{% top %} diff --git a/docs/internals/ide_setup.md b/docs/internals/ide_setup.md index 31ad6b8d8e847..02d54e7810497 100644 --- a/docs/internals/ide_setup.md +++ b/docs/internals/ide_setup.md @@ -119,3 +119,5 @@ due to deficiencies of the old Eclipse version bundled with Scala IDE 3.0.3 or due to version incompatibilities with the bundled Scala version in Scala IDE 4.4.1. **We recommend to use IntelliJ instead (see [above](#intellij-idea))** + +{% top %} diff --git a/docs/internals/job_scheduling.md b/docs/internals/job_scheduling.md index 74062c68e24f4..668dfa3f42d8e 100644 --- a/docs/internals/job_scheduling.md +++ b/docs/internals/job_scheduling.md @@ -101,3 +101,5 @@ For that reason, the execution of an ExecutionVertex is tracked in an {% gh_link
States and Transitions of Task Executions
+ +{% top %} diff --git a/docs/internals/stream_checkpointing.md b/docs/internals/stream_checkpointing.md index 330b0aae3f4ad..8fc96cc647dda 100644 --- a/docs/internals/stream_checkpointing.md +++ b/docs/internals/stream_checkpointing.md @@ -169,3 +169,5 @@ Operators that checkpoint purely synchronously return an already completed `Futu If an asynchronous operation needs to be performed, it is executed in the `run()` method of that `FutureTask`. The tasks are cancelable, so that streams and other resource consuming handles can be released. + +{% top %} diff --git a/docs/internals/task_lifecycle.md b/docs/internals/task_lifecycle.md index fed2cb9792f17..182c99c9a9813 100644 --- a/docs/internals/task_lifecycle.md +++ b/docs/internals/task_lifecycle.md @@ -190,3 +190,5 @@ In the previous sections we described the lifecycle of a task that runs till com at any point, then the normal execution is interrupted and the only operations performed from that point on are the timer service shutdown, the task-specific cleanup, the disposal of the operators, and the general task cleanup, as described above. + +{% top %} diff --git a/docs/monitoring/application_profiling.md b/docs/monitoring/application_profiling.md index 65ef45e4fa5fb..721bc31a2d235 100644 --- a/docs/monitoring/application_profiling.md +++ b/docs/monitoring/application_profiling.md @@ -52,3 +52,5 @@ compiler used to inspect inlining decisions, hot methods, bytecode, and assembly ~~~ env.java.opts: "-XX:+UnlockDiagnosticVMOptions -XX:+TraceClassLoading -XX:+LogCompilation -XX:LogFile=${FLINK_LOG_PREFIX}.jit -XX:+PrintAssembly" ~~~ + +{% top %} diff --git a/docs/monitoring/back_pressure.md b/docs/monitoring/back_pressure.md index f047066831c97..28a4dd20623c1 100644 --- a/docs/monitoring/back_pressure.md +++ b/docs/monitoring/back_pressure.md @@ -79,3 +79,5 @@ If you see status **OK** for the tasks, there is no indication of back pressure. + +{% top %} diff --git a/docs/monitoring/checkpoint_monitoring.md b/docs/monitoring/checkpoint_monitoring.md index 7cf06d118f9b3..6c2c2899a3953 100644 --- a/docs/monitoring/checkpoint_monitoring.md +++ b/docs/monitoring/checkpoint_monitoring.md @@ -113,3 +113,5 @@ When you click on a *More details* link for a checkpoint, you get a Minumum/Aver
Checkpoint Monitoring: Subtasks
+ +{% top %} diff --git a/docs/monitoring/debugging_classloading.md b/docs/monitoring/debugging_classloading.md index c072777f37050..4f57c10aacc29 100644 --- a/docs/monitoring/debugging_classloading.md +++ b/docs/monitoring/debugging_classloading.md @@ -140,3 +140,4 @@ This documentation page explains [relocating classes using the shade plugin](htt Note that some of Flink's dependencies, such as `guava` are shaded away by the maintainers of Flink, so users usually don't have to worry about it. +{% top %} diff --git a/docs/monitoring/debugging_event_time.md b/docs/monitoring/debugging_event_time.md index 8355b62e7677d..10a3fb20fdb26 100644 --- a/docs/monitoring/debugging_event_time.md +++ b/docs/monitoring/debugging_event_time.md @@ -54,3 +54,4 @@ For local setups, we recommend using the JMX metric reporter and a tool like [Vi - Approach 1: Watermark stays late (indicated completeness), windows fire early - Approach 2: Watermark heuristic with maximum lateness, windows accept late data +{% top %} diff --git a/docs/monitoring/historyserver.md b/docs/monitoring/historyserver.md index e109512a5804f..61660a5c59ba8 100644 --- a/docs/monitoring/historyserver.md +++ b/docs/monitoring/historyserver.md @@ -95,3 +95,5 @@ Values in angle brackets are variables, for example `http://hostname:port/jobs/< - `/jobs//vertices//subtasks//attempts/` - `/jobs//vertices//subtasks//attempts//accumulators` - `/jobs//plan` + +{% top %} diff --git a/docs/monitoring/logging.md b/docs/monitoring/logging.md index ee6a316c55c28..b548d416693ee 100644 --- a/docs/monitoring/logging.md +++ b/docs/monitoring/logging.md @@ -96,3 +96,5 @@ catch(Exception exception){ LOG.error("An {} occurred.", "error", exception); } ~~~ + +{% top %} diff --git a/docs/monitoring/rest_api.md b/docs/monitoring/rest_api.md index c5efcc214a083..4202886587f57 100644 --- a/docs/monitoring/rest_api.md +++ b/docs/monitoring/rest_api.md @@ -700,3 +700,5 @@ Response: ~~~ {"jobid": "869a9868d49c679e7355700e0857af85"} ~~~ + +{% top %} diff --git a/docs/ops/cli.md b/docs/ops/cli.md index 7b36177223f39..11c8caf8ba1ef 100644 --- a/docs/ops/cli.md +++ b/docs/ops/cli.md @@ -353,3 +353,5 @@ Action "savepoint" triggers savepoints for a running job or disposes existing on Options for yarn-cluster mode: -yid,--yarnapplicationId Attach to running YARN session ~~~ + +{% top %} diff --git a/docs/ops/config.md b/docs/ops/config.md index 9d2405e525e19..64ef48fb838a0 100644 --- a/docs/ops/config.md +++ b/docs/ops/config.md @@ -723,3 +723,5 @@ Each Flink TaskManager provides processing slots in the cluster. The number of s When starting a Flink application, users can supply the default number of slots to use for that job. The command line value therefore is called `-p` (for parallelism). In addition, it is possible to [set the number of slots in the programming APIs]({{site.baseurl}}/dev/parallel.html) for the whole application and for individual operators. + +{% top %} diff --git a/docs/ops/deployment/aws.md b/docs/ops/deployment/aws.md index 9c6e302ffeddc..1a05bfdda6b0d 100644 --- a/docs/ops/deployment/aws.md +++ b/docs/ops/deployment/aws.md @@ -372,3 +372,5 @@ o.a.f.runtime.fs.hdfs.HadoopFileSystem.create(HadoopFileSystem.java:404) at o.a.f.runtime.fs.hdfs.HadoopFileSystem.create(HadoopFileSystem.java:48) at ... 25 more ``` + +{% top %} diff --git a/docs/ops/deployment/gce_setup.md b/docs/ops/deployment/gce_setup.md index 2925737a34374..0b9482cdd8bcc 100644 --- a/docs/ops/deployment/gce_setup.md +++ b/docs/ops/deployment/gce_setup.md @@ -91,3 +91,5 @@ To bring up the Flink cluster on Google Compute Engine, execute: Shutting down a cluster is as simple as executing ./bdutil -e extensions/flink/flink_env.sh delete + +{% top %} diff --git a/docs/ops/deployment/mapr_setup.md b/docs/ops/deployment/mapr_setup.md index 7575bdc8c1c75..19920ad9076b2 100644 --- a/docs/ops/deployment/mapr_setup.md +++ b/docs/ops/deployment/mapr_setup.md @@ -130,3 +130,5 @@ java.lang.Exception: unable to establish the security context Caused by: o.a.f.r.security.modules.SecurityModule$SecurityInstallException: Unable to set the Hadoop login user Caused by: java.io.IOException: failure to login: Unable to obtain MapR credentials ``` + +{% top %} diff --git a/docs/ops/deployment/mesos.md b/docs/ops/deployment/mesos.md index 2fa340d6d6d30..36b6df7ffdb42 100644 --- a/docs/ops/deployment/mesos.md +++ b/docs/ops/deployment/mesos.md @@ -267,3 +267,5 @@ May be set to -1 to disable this feature. `mesos.resourcemanager.tasks.hostname`: Optional value to define the TaskManager's hostname. The pattern `_TASK_` is replaced by the actual id of the Mesos task. This can be used to configure the TaskManager to use Mesos DNS (e.g. `_TASK_.flink-service.mesos`) for name lookups. (**NO DEFAULT**) `mesos.resourcemanager.tasks.bootstrap-cmd`: A command which is executed before the TaskManager is started (**NO DEFAULT**). + +{% top %} diff --git a/docs/ops/deployment/yarn_setup.md b/docs/ops/deployment/yarn_setup.md index 8c435f7cacda6..0fb5bf6584d85 100644 --- a/docs/ops/deployment/yarn_setup.md +++ b/docs/ops/deployment/yarn_setup.md @@ -336,3 +336,5 @@ The next step of the client is to request (step 2) a YARN container to start the The *JobManager* and AM are running in the same container. Once they successfully started, the AM knows the address of the JobManager (its own host). It is generating a new Flink configuration file for the TaskManagers (so that they can connect to the JobManager). The file is also uploaded to HDFS. Additionally, the *AM* container is also serving Flink's web interface. All ports the YARN code is allocating are *ephemeral ports*. This allows users to execute multiple Flink YARN sessions in parallel. After that, the AM starts allocating the containers for Flink's TaskManagers, which will download the jar file and the modified configuration from the HDFS. Once these steps are completed, Flink is set up and ready to accept Jobs. + +{% top %} diff --git a/docs/ops/jobmanager_high_availability.md b/docs/ops/jobmanager_high_availability.md index 7dd7d4ca63b0d..e73353ba81ea6 100644 --- a/docs/ops/jobmanager_high_availability.md +++ b/docs/ops/jobmanager_high_availability.md @@ -237,3 +237,5 @@ server.Y=addressY:peerPort:leaderPort The script `bin/start-zookeeper-quorum.sh` will start a ZooKeeper server on each of the configured hosts. The started processes start ZooKeeper servers via a Flink wrapper, which reads the configuration from `conf/zoo.cfg` and makes sure to set some required configuration values for convenience. In production setups, it is recommended to manage your own ZooKeeper installation. + +{% top %} diff --git a/docs/ops/production_ready.md b/docs/ops/production_ready.md index c58ce5b794289..303e7a71bb6cc 100644 --- a/docs/ops/production_ready.md +++ b/docs/ops/production_ready.md @@ -86,3 +86,5 @@ stream processing. However, RocksDB can have worse performance than, for example you are sure that your state will never exceed main memory and blocking the stream processing to write it is not an issue, you **could consider** to not use the RocksDB backends. However, at this point, we **strongly recommend** using RocksDB for production. + +{% top %} diff --git a/docs/ops/security-kerberos.md b/docs/ops/security-kerberos.md index eac72f1cef0f5..5589057a3276b 100644 --- a/docs/ops/security-kerberos.md +++ b/docs/ops/security-kerberos.md @@ -117,3 +117,5 @@ Steps to run a secure Flink cluster using `kinit`: Each component that uses Kerberos is independently responsible for renewing the Kerberos ticket-granting-ticket (TGT). Hadoop, ZooKeeper, and Kafka all renew the TGT automatically when provided a keytab. In the delegation token scenario, YARN itself renews the token (up to its maximum lifespan). + +{% top %} diff --git a/docs/ops/security-ssl.md b/docs/ops/security-ssl.md index 7c7268a9ab342..8c7bf2bdbd324 100644 --- a/docs/ops/security-ssl.md +++ b/docs/ops/security-ssl.md @@ -142,3 +142,4 @@ flink run -m yarn-cluster -yt deploy-keys/ TestJob.jar When deployed using YARN, flink's web dashboard is accessible through YARN proxy's Tracking URL. To ensure that the YARN proxy is able to access flink's https url you need to configure YARN proxy to accept flink's SSL certificates. Add the custom CA certificate into Java's default trustore on the YARN Proxy node. +{% top %} diff --git a/docs/ops/state/checkpoints.md b/docs/ops/state/checkpoints.md index 96c7a2084ae21..690680f0f9035 100644 --- a/docs/ops/state/checkpoints.md +++ b/docs/ops/state/checkpoints.md @@ -99,3 +99,5 @@ above). ```sh $ bin/flink run -s :checkpointMetaDataPath [:runArgs] ``` + +{% top %} diff --git a/docs/ops/state/large_state_tuning.md b/docs/ops/state/large_state_tuning.md index aa673a4170c0b..dd3e404845d60 100644 --- a/docs/ops/state/large_state_tuning.md +++ b/docs/ops/state/large_state_tuning.md @@ -235,3 +235,5 @@ Compression can be activated through the `ExecutionConfig`: **Notice:** The compression option has no impact on incremental snapshots, because they are using RocksDB's internal format which is always using snappy compression out of the box. + +{% top %} diff --git a/docs/ops/state/savepoints.md b/docs/ops/state/savepoints.md index 1d82d2b10d3e5..d6d4c53df3d07 100644 --- a/docs/ops/state/savepoints.md +++ b/docs/ops/state/savepoints.md @@ -196,3 +196,5 @@ If you did not assign IDs, the auto generated IDs of the stateful operators will If the savepoint was triggered with Flink >= 1.2.0 and using no deprecated state API like `Checkpointed`, you can simply restore the program from a savepoint and specify a new parallelism. If you are resuming from a savepoint triggered with Flink < 1.2.0 or using now deprecated APIs you first have to migrate your job and savepoint to Flink >= 1.2.0 before being able to change the parallelism. See the [upgrading jobs and Flink versions guide]({{ site.baseurl }}/ops/upgrading.html). + +{% top %} diff --git a/docs/ops/state/state_backends.md b/docs/ops/state/state_backends.md index b53bceff4247d..422df3ec509b0 100644 --- a/docs/ops/state/state_backends.md +++ b/docs/ops/state/state_backends.md @@ -167,3 +167,5 @@ state.backend: filesystem state.backend.fs.checkpointdir: hdfs://namenode:40010/flink/checkpoints ~~~ + +{% top %} diff --git a/docs/ops/upgrading.md b/docs/ops/upgrading.md index 12d15ea0c399c..2a34c172ae148 100644 --- a/docs/ops/upgrading.md +++ b/docs/ops/upgrading.md @@ -240,3 +240,5 @@ Savepoints are compatible across Flink versions as indicated by the table below: + +{% top %} diff --git a/docs/quickstart/java_api_quickstart.md b/docs/quickstart/java_api_quickstart.md index c21e06e3df487..109240b932500 100644 --- a/docs/quickstart/java_api_quickstart.md +++ b/docs/quickstart/java_api_quickstart.md @@ -192,3 +192,5 @@ public static final class LineSplitter implements FlatMapFunctionExample Job View This concludes our little tour of Flink. If you have any questions, please don't hesitate to ask on our [Mailing Lists](http://flink.apache.org/community.html#mailing-lists). + +{% top %} diff --git a/docs/quickstart/scala_api_quickstart.md b/docs/quickstart/scala_api_quickstart.md index 9e563ed25eb2c..33013e4b745fc 100644 --- a/docs/quickstart/scala_api_quickstart.md +++ b/docs/quickstart/scala_api_quickstart.md @@ -260,3 +260,5 @@ For a complete overview over our API, have a look at the sections. If you have any trouble, ask on our [Mailing List](http://mail-archives.apache.org/mod_mbox/flink-dev/). We are happy to provide help. + +{% top %} diff --git a/docs/quickstart/setup_quickstart.md b/docs/quickstart/setup_quickstart.md index 0e4f3d8b3164a..3d40ddf2b600a 100644 --- a/docs/quickstart/setup_quickstart.md +++ b/docs/quickstart/setup_quickstart.md @@ -299,3 +299,5 @@ window of processing time, as long as words are floating in. ## Next Steps Check out some more [examples]({{ site.baseurl }}/examples) to get a better feel for Flink's programming APIs. When you are done with that, go ahead and read the [streaming guide]({{ site.baseurl }}/dev/datastream_api.html). + +{% top %} diff --git a/docs/search-results.md b/docs/search-results.md index 2c37b4461cbe3..5d8de997a2281 100644 --- a/docs/search-results.md +++ b/docs/search-results.md @@ -32,3 +32,5 @@ under the License. + +{% top %}