spark-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
Subject spark git commit: [STREAMING][KAFKA][DOC] clarify kafka settings needed for larger batches
Date Fri, 21 Oct 2016 23:27:22 GMT
Repository: spark
Updated Branches:
  refs/heads/master 268ccb9a4 -> c9720b219

[STREAMING][KAFKA][DOC] clarify kafka settings needed for larger batches

## What changes were proposed in this pull request?

Minor doc change to mention kafka configuration for larger spark batches.

## How was this patch tested?

Doc change only, confirmed via jekyll.

The configuration issue was discussed / confirmed with users on the mailing list.

Author: cody koeninger <>

Closes #15570 from koeninger/kafka-doc-heartbeat.


Branch: refs/heads/master
Commit: c9720b2195a465653690b3e221ce789142217b0d
Parents: 268ccb9
Author: cody koeninger <>
Authored: Fri Oct 21 16:27:19 2016 -0700
Committer: Shixiong Zhu <>
Committed: Fri Oct 21 16:27:19 2016 -0700

 docs/ | 1 +
 1 file changed, 1 insertion(+)
diff --git a/docs/ b/docs/
index 456b845..de95ea9 100644
--- a/docs/
+++ b/docs/
@@ -48,6 +48,7 @@ Each item in the stream is a [ConsumerRecord](
 For possible kafkaParams, see [Kafka consumer config docs](
+If your Spark batch duration is larger than the default Kafka heartbeat session timeout (30
seconds), increase and appropriately.  For batches
larger than 5 minutes, this will require changing on the broker.
 Note that the example sets to false, for discussion see [Storing Offsets](streaming-kafka-0-10-integration.html#storing-offsets)
 ### LocationStrategies

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message