beam-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Joshua Fox (JIRA)" <>
Subject [jira] [Commented] (BEAM-991) DatastoreIO Write should flush early for large batches
Date Thu, 01 Dec 2016 11:23:58 GMT


Joshua Fox commented on BEAM-991:

Daniel Halperin, unfortunately I cannot participate too much in the development and full testing,
both for the usual  reason of time commitments but also because I am not close enough to the
infrastructure to fully understand what is happening, particularly in edge cases. However,
I will be glad to serve as an early user.

My situation: I wrote a Datastore backup in Dataflow, after Google's Backup and Managed Backup
tool failed on various bugs.

However, this new tool cannot copy any Kind with Item size >20 KB (max in Datastore is
1 MB). So, I can only use it to backup Kinds with average Item size 10 KB (since there may
be variation.) I wrote a second a simple multithreaded but nondistributed backup tool, using
 "select-insert"  loops. It works, but of course is less scalable and more expensive than
the Dataflow tool. I use a combination of the Dataflow and the nondistributed tool now, for
different Kinds, but can easily switch to using just Dataflow.

> DatastoreIO Write should flush early for large batches
> ------------------------------------------------------
>                 Key: BEAM-991
>                 URL:
>             Project: Beam
>          Issue Type: Bug
>          Components: sdk-java-gcp
>            Reporter: Vikas Kedigehalli
>            Assignee: Vikas Kedigehalli
> If entities are large (avg size > 20KB) then the a single batched write (500 entities)
would exceed the Datastore size limit of a single request (10MB) from
> First reported in:

This message was sent by Atlassian JIRA

View raw message