From commits-return-102519-archive-asf-public=cust-asf.ponee.io@beam.apache.org Tue Jul 9 09:15:29 2019 Return-Path: X-Original-To: archive-asf-public@cust-asf.ponee.io Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [207.244.88.153]) by mx-eu-01.ponee.io (Postfix) with SMTP id A7D3E18062B for ; Tue, 9 Jul 2019 11:15:28 +0200 (CEST) Received: (qmail 92673 invoked by uid 500); 9 Jul 2019 09:15:27 -0000 Mailing-List: contact commits-help@beam.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@beam.apache.org Delivered-To: mailing list commits@beam.apache.org Received: (qmail 92664 invoked by uid 99); 9 Jul 2019 09:15:27 -0000 Received: from ec2-52-202-80-70.compute-1.amazonaws.com (HELO gitbox.apache.org) (52.202.80.70) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 09 Jul 2019 09:15:27 +0000 Received: by gitbox.apache.org (ASF Mail Server at gitbox.apache.org, from userid 33) id 90E2A87AD7; Tue, 9 Jul 2019 09:15:27 +0000 (UTC) Date: Tue, 09 Jul 2019 09:15:26 +0000 To: "commits@beam.apache.org" Subject: [beam] branch master updated: Consider Elasticsearch as one word in camelCase. MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 8bit Message-ID: <156266372433.3672.6675816368450559550@gitbox.apache.org> From: echauchot@apache.org X-Git-Host: gitbox.apache.org X-Git-Repo: beam X-Git-Refname: refs/heads/master X-Git-Reftype: branch X-Git-Oldrev: 69bb363334fdfe7f812bbb2a0bc210419be0409d X-Git-Newrev: ef4a51d3900d641204551b56bb88def11c05e832 X-Git-Rev: ef4a51d3900d641204551b56bb88def11c05e832 X-Git-NotificationType: ref_changed_plus_diff X-Git-Multimail-Version: 1.5.dev Auto-Submitted: auto-generated This is an automated email from the ASF dual-hosted git repository. echauchot pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/beam.git The following commit(s) were added to refs/heads/master by this push: new ef4a51d Consider Elasticsearch as one word in camelCase. new c2343c8 Merge pull request #9008 from RyanSkraba/BEAM-7698-rename-elasticsearch ef4a51d is described below commit ef4a51d3900d641204551b56bb88def11c05e832 Author: Ryan Skraba AuthorDate: Tue Jul 2 18:43:00 2019 +0200 Consider Elasticsearch as one word in camelCase. --- .../sdk/io/elasticsearch/ElasticsearchIOIT.java | 6 +-- .../sdk/io/elasticsearch/ElasticsearchIOTest.java | 2 +- .../sdk/io/elasticsearch/ElasticsearchIOIT.java | 8 +-- .../sdk/io/elasticsearch/ElasticsearchIOIT.java | 8 +-- .../io/elasticsearch/ElasticsearchIOITCommon.java | 2 +- .../elasticsearch/ElasticsearchIOTestCommon.java | 58 +++++++++++----------- ...estUtils.java => ElasticsearchIOTestUtils.java} | 6 +-- .../beam/sdk/io/elasticsearch/ElasticsearchIO.java | 4 +- website/src/_posts/2018-01-09-beam-a-look-back.md | 2 +- website/src/documentation/io/built-in-hadoop.md | 16 +++--- 10 files changed, 56 insertions(+), 56 deletions(-) diff --git a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-2/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-2/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java index 9eccb7f..c0ea12c 100644 --- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-2/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java +++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-2/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java @@ -79,8 +79,8 @@ public class ElasticsearchIOIT { @AfterClass public static void afterClass() throws Exception { - ElasticSearchIOTestUtils.deleteIndex(writeConnectionConfiguration, restClient); - ElasticSearchIOTestUtils.deleteIndex(updateConnectionConfiguration, restClient); + ElasticsearchIOTestUtils.deleteIndex(writeConnectionConfiguration, restClient); + ElasticsearchIOTestUtils.deleteIndex(updateConnectionConfiguration, restClient); restClient.close(); } @@ -130,7 +130,7 @@ public class ElasticsearchIOIT { */ @Test public void testWritePartialUpdate() throws Exception { - ElasticSearchIOTestUtils.copyIndex( + ElasticsearchIOTestUtils.copyIndex( restClient, readConnectionConfiguration.getIndex(), updateConnectionConfiguration.getIndex()); diff --git a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-2/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTest.java b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-2/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTest.java index 19bda80..4fe5844 100644 --- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-2/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTest.java +++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-2/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTest.java @@ -116,7 +116,7 @@ public class ElasticsearchIOTest implements Serializable { @Before public void before() throws Exception { - ElasticSearchIOTestUtils.deleteIndex(connectionConfiguration, restClient); + ElasticsearchIOTestUtils.deleteIndex(connectionConfiguration, restClient); } @Test diff --git a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-5/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-5/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java index 8fdb444..cf52282 100644 --- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-5/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java +++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-5/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java @@ -82,8 +82,8 @@ public class ElasticsearchIOIT { @AfterClass public static void afterClass() throws Exception { - ElasticSearchIOTestUtils.deleteIndex(writeConnectionConfiguration, restClient); - ElasticSearchIOTestUtils.deleteIndex(updateConnectionConfiguration, restClient); + ElasticsearchIOTestUtils.deleteIndex(writeConnectionConfiguration, restClient); + ElasticsearchIOTestUtils.deleteIndex(updateConnectionConfiguration, restClient); restClient.close(); } @@ -130,11 +130,11 @@ public class ElasticsearchIOIT { /** * This test verifies volume partial updates of Elasticsearch. The test dataset index is cloned * and then a new field is added to each document using a partial update. The test then asserts - * the updates where appied. + * the updates were applied. */ @Test public void testWritePartialUpdate() throws Exception { - ElasticSearchIOTestUtils.copyIndex( + ElasticsearchIOTestUtils.copyIndex( restClient, readConnectionConfiguration.getIndex(), updateConnectionConfiguration.getIndex()); diff --git a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-6/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-6/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java index b89bac4..9629440 100644 --- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-6/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java +++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-6/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOIT.java @@ -82,8 +82,8 @@ public class ElasticsearchIOIT { @AfterClass public static void afterClass() throws Exception { - ElasticSearchIOTestUtils.deleteIndex(writeConnectionConfiguration, restClient); - ElasticSearchIOTestUtils.deleteIndex(updateConnectionConfiguration, restClient); + ElasticsearchIOTestUtils.deleteIndex(writeConnectionConfiguration, restClient); + ElasticsearchIOTestUtils.deleteIndex(updateConnectionConfiguration, restClient); restClient.close(); } @@ -130,11 +130,11 @@ public class ElasticsearchIOIT { /** * This test verifies volume partial updates of Elasticsearch. The test dataset index is cloned * and then a new field is added to each document using a partial update. The test then asserts - * the updates where appied. + * the updates were applied. */ @Test public void testWritePartialUpdate() throws Exception { - ElasticSearchIOTestUtils.copyIndex( + ElasticsearchIOTestUtils.copyIndex( restClient, readConnectionConfiguration.getIndex(), updateConnectionConfiguration.getIndex()); diff --git a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOITCommon.java b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOITCommon.java index 6ef38bd..76b88f8 100644 --- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOITCommon.java +++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOITCommon.java @@ -99,7 +99,7 @@ public class ElasticsearchIOITCommon { ConnectionConfiguration connectionConfiguration = getConnectionConfiguration(options, IndexMode.READ); try (RestClient restClient = connectionConfiguration.createClient()) { - ElasticSearchIOTestUtils.insertTestDocuments( + ElasticsearchIOTestUtils.insertTestDocuments( connectionConfiguration, NUM_DOCS_ITESTS, restClient); } } diff --git a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTestCommon.java b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTestCommon.java index 112e0e2..90ca521 100644 --- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTestCommon.java +++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTestCommon.java @@ -17,16 +17,16 @@ */ package org.apache.beam.sdk.io.elasticsearch; -import static org.apache.beam.sdk.io.elasticsearch.ElasticSearchIOTestUtils.FAMOUS_SCIENTISTS; -import static org.apache.beam.sdk.io.elasticsearch.ElasticSearchIOTestUtils.NUM_SCIENTISTS; -import static org.apache.beam.sdk.io.elasticsearch.ElasticSearchIOTestUtils.countByMatch; -import static org.apache.beam.sdk.io.elasticsearch.ElasticSearchIOTestUtils.countByScientistName; -import static org.apache.beam.sdk.io.elasticsearch.ElasticSearchIOTestUtils.refreshIndexAndGetCurrentNumDocs; import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.BoundedElasticsearchSource; import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.ConnectionConfiguration; import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Read; import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.RetryConfiguration.DEFAULT_RETRY_PREDICATE; import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIO.Write; +import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTestUtils.FAMOUS_SCIENTISTS; +import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTestUtils.NUM_SCIENTISTS; +import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTestUtils.countByMatch; +import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTestUtils.countByScientistName; +import static org.apache.beam.sdk.io.elasticsearch.ElasticsearchIOTestUtils.refreshIndexAndGetCurrentNumDocs; import static org.apache.beam.sdk.testing.SourceTestUtils.readFromSource; import static org.hamcrest.Matchers.greaterThan; import static org.hamcrest.Matchers.lessThan; @@ -129,7 +129,7 @@ class ElasticsearchIOTestCommon implements Serializable { void testSplit(final int desiredBundleSizeBytes) throws Exception { if (!useAsITests) { - ElasticSearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); + ElasticsearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); } PipelineOptions options = PipelineOptionsFactory.create(); Read read = ElasticsearchIO.read().withConnectionConfiguration(connectionConfiguration); @@ -166,7 +166,7 @@ class ElasticsearchIOTestCommon implements Serializable { void testSizes() throws Exception { if (!useAsITests) { - ElasticSearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); + ElasticsearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); } PipelineOptions options = PipelineOptionsFactory.create(); Read read = ElasticsearchIO.read().withConnectionConfiguration(connectionConfiguration); @@ -181,7 +181,7 @@ class ElasticsearchIOTestCommon implements Serializable { void testRead() throws Exception { if (!useAsITests) { - ElasticSearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); + ElasticsearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); } PCollection output = @@ -198,7 +198,7 @@ class ElasticsearchIOTestCommon implements Serializable { void testReadWithQuery() throws Exception { if (!useAsITests) { - ElasticSearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); + ElasticsearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); } String query = @@ -225,7 +225,7 @@ class ElasticsearchIOTestCommon implements Serializable { /** Test reading metadata by reading back the id of a document after writing it. */ void testReadWithMetadata() throws Exception { if (!useAsITests) { - ElasticSearchIOTestUtils.insertTestDocuments(connectionConfiguration, 1, restClient); + ElasticsearchIOTestUtils.insertTestDocuments(connectionConfiguration, 1, restClient); } PCollection output = @@ -248,8 +248,8 @@ class ElasticsearchIOTestCommon implements Serializable { .withConnectionConfiguration(connectionConfiguration) .withMaxBatchSize(BATCH_SIZE); List input = - ElasticSearchIOTestUtils.createDocuments( - numDocs, ElasticSearchIOTestUtils.InjectionMode.INJECT_SOME_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + numDocs, ElasticsearchIOTestUtils.InjectionMode.INJECT_SOME_INVALID_DOCS); expectedException.expect(isA(IOException.class)); expectedException.expectMessage( new CustomMatcher("RegExp matcher") { @@ -286,8 +286,8 @@ class ElasticsearchIOTestCommon implements Serializable { // so we test the Writer as a DoFn outside of a runner. try (DoFnTester fnTester = DoFnTester.of(new Write.WriteFn(write))) { List input = - ElasticSearchIOTestUtils.createDocuments( - numDocs, ElasticSearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + numDocs, ElasticsearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); long numDocsProcessed = 0; long numDocsInserted = 0; for (String document : input) { @@ -327,8 +327,8 @@ class ElasticsearchIOTestCommon implements Serializable { // so we test the Writer as a DoFn outside of a runner. try (DoFnTester fnTester = DoFnTester.of(new Write.WriteFn(write))) { List input = - ElasticSearchIOTestUtils.createDocuments( - numDocs, ElasticSearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + numDocs, ElasticsearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); long numDocsProcessed = 0; long sizeProcessed = 0; long numDocsInserted = 0; @@ -383,8 +383,8 @@ class ElasticsearchIOTestCommon implements Serializable { */ void testWriteWithIdFn() throws Exception { List data = - ElasticSearchIOTestUtils.createDocuments( - numDocs, ElasticSearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + numDocs, ElasticsearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); pipeline .apply(Create.of(data)) .apply( @@ -413,8 +413,8 @@ class ElasticsearchIOTestCommon implements Serializable { long adjustedNumDocs = docsPerScientist * FAMOUS_SCIENTISTS.length; List data = - ElasticSearchIOTestUtils.createDocuments( - adjustedNumDocs, ElasticSearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + adjustedNumDocs, ElasticsearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); pipeline .apply(Create.of(data)) .apply( @@ -459,8 +459,8 @@ class ElasticsearchIOTestCommon implements Serializable { long adjustedNumDocs = (numDocs & 1) == 0 ? numDocs : numDocs + 1; List data = - ElasticSearchIOTestUtils.createDocuments( - adjustedNumDocs, ElasticSearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + adjustedNumDocs, ElasticsearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); pipeline .apply(Create.of(data)) .apply( @@ -485,8 +485,8 @@ class ElasticsearchIOTestCommon implements Serializable { */ void testWriteWithFullAddressing() throws Exception { List data = - ElasticSearchIOTestUtils.createDocuments( - numDocs, ElasticSearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + numDocs, ElasticsearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); pipeline .apply(Create.of(data)) .apply( @@ -514,7 +514,7 @@ class ElasticsearchIOTestCommon implements Serializable { */ void testWritePartialUpdate() throws Exception { if (!useAsITests) { - ElasticSearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); + ElasticsearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); } // defensive coding to ensure our initial state is as expected @@ -553,10 +553,10 @@ class ElasticsearchIOTestCommon implements Serializable { /** Tests partial updates with errors by adding some invalid info to test set. */ void testWritePartialUpdateWithErrors() throws Exception { // put a mapping to simulate error of insertion - ElasticSearchIOTestUtils.setIndexMapping(connectionConfiguration, restClient); + ElasticsearchIOTestUtils.setIndexMapping(connectionConfiguration, restClient); if (!useAsITests) { - ElasticSearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); + ElasticsearchIOTestUtils.insertTestDocuments(connectionConfiguration, numDocs, restClient); } // try to partial update a document with an incompatible date format for the age to generate @@ -660,8 +660,8 @@ class ElasticsearchIOTestCommon implements Serializable { private void executeWriteTest(ElasticsearchIO.Write write) throws Exception { List data = - ElasticSearchIOTestUtils.createDocuments( - numDocs, ElasticSearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + numDocs, ElasticsearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); pipeline.apply(Create.of(data)).apply(write); pipeline.run(); diff --git a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticSearchIOTestUtils.java b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTestUtils.java similarity index 98% rename from sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticSearchIOTestUtils.java rename to sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTestUtils.java index eacbbc5..ee9e47c 100644 --- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticSearchIOTestUtils.java +++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/src/test/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIOTestUtils.java @@ -34,7 +34,7 @@ import org.elasticsearch.client.Response; import org.elasticsearch.client.RestClient; /** Test utilities to use with {@link ElasticsearchIO}. */ -class ElasticSearchIOTestUtils { +class ElasticsearchIOTestUtils { static final String[] FAMOUS_SCIENTISTS = { "Einstein", "Darwin", @@ -100,8 +100,8 @@ class ElasticSearchIOTestUtils { ConnectionConfiguration connectionConfiguration, long numDocs, RestClient restClient) throws IOException { List data = - ElasticSearchIOTestUtils.createDocuments( - numDocs, ElasticSearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); + ElasticsearchIOTestUtils.createDocuments( + numDocs, ElasticsearchIOTestUtils.InjectionMode.DO_NOT_INJECT_INVALID_DOCS); StringBuilder bulkRequest = new StringBuilder(); int i = 0; for (String document : data) { diff --git a/sdks/java/io/elasticsearch/src/main/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIO.java b/sdks/java/io/elasticsearch/src/main/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIO.java index ec688fb..1d1ba34 100644 --- a/sdks/java/io/elasticsearch/src/main/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIO.java +++ b/sdks/java/io/elasticsearch/src/main/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIO.java @@ -1013,7 +1013,7 @@ public class ElasticsearchIO { * docs (like Elasticsearch bulk size advice). See * https://www.elastic.co/guide/en/elasticsearch/guide/current/bulk.html Depending on the * execution engine, size of bundles may vary, this sets the maximum size. Change this if you - * need to have smaller ElasticSearch bulks. + * need to have smaller Elasticsearch bulks. * * @param batchSize maximum batch size in number of documents * @return the {@link Write} with connection batch size set @@ -1029,7 +1029,7 @@ public class ElasticsearchIO { * (like Elasticsearch bulk size advice). See * https://www.elastic.co/guide/en/elasticsearch/guide/current/bulk.html Depending on the * execution engine, size of bundles may vary, this sets the maximum size. Change this if you - * need to have smaller ElasticSearch bulks. + * need to have smaller Elasticsearch bulks. * * @param batchSizeBytes maximum batch size in bytes * @return the {@link Write} with connection batch size in bytes set diff --git a/website/src/_posts/2018-01-09-beam-a-look-back.md b/website/src/_posts/2018-01-09-beam-a-look-back.md index 90c16b0..5c58bc9 100644 --- a/website/src/_posts/2018-01-09-beam-a-look-back.md +++ b/website/src/_posts/2018-01-09-beam-a-look-back.md @@ -64,7 +64,7 @@ new and updated runners were developed: In addition to runners, Beam added new IO connectors, some notable ones being the Cassandra, MQTT, AMQP, HBase/HCatalog, JDBC, Solr, Tika, Redis, and -ElasticSearch connectors. Beam’s IO connectors make it possible to read from or +Elasticsearch connectors. Beam’s IO connectors make it possible to read from or write to data sources/sinks even when they are not natively supported by the underlying execution engine. Beam also provides fully pluggable filesystem support, allowing us to support and extend our coverage to HDFS, S3, Azure diff --git a/website/src/documentation/io/built-in-hadoop.md b/website/src/documentation/io/built-in-hadoop.md index fd330ec..09fcd7f 100644 --- a/website/src/documentation/io/built-in-hadoop.md +++ b/website/src/documentation/io/built-in-hadoop.md @@ -186,13 +186,13 @@ SimpleFunction cassandraOutputValueType = SimpleFunction> elasticData = p.apply("read", - HadoopFormatIO.read().withConfiguration(elasticSearchConf)); + HadoopFormatIO.read().withConfiguration(elasticsearchConf)); ``` ```py