beam-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: beam_PostCommit_Python_Verify #2315
Date Thu, 25 May 2017 16:05:00 GMT
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/2315/display/redirect>

------------------------------------------
[...truncated 576.69 KB...]
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "kind:windowed_value", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": [
                        {
                          "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                          "component_encodings": []
                        }, 
                        {
                          "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                          "component_encodings": []
                        }
                      ], 
                      "is_pair_like": true
                    }, 
                    {
                      "@type": "kind:global_window"
                    }
                  ], 
                  "is_wrapper": true
                }
              ]
            }, 
            "output_name": "out", 
            "user_name": "write/Write/WriteImpl/FinalizeWrite/SideInput-s16.output"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s14"
        }, 
        "user_name": "write/Write/WriteImpl/FinalizeWrite/SideInput-s16"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s17", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_finalize_write"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {
          "SideInput-s15": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "SideInput-s15"
          }, 
          "SideInput-s16": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "SideInput-s16"
          }
        }, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "write/Write/WriteImpl/FinalizeWrite.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s7"
        }, 
        "serialized_fn": "<string of 1056 bytes>", 
        "user_name": "write/Write/WriteImpl/FinalizeWrite/Do"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2017-05-25T15:59:18.434829Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2017-05-25_08_59_17-15727091547768182389'
 location: u'global'
 name: u'beamapp-jenkins-0525155916-528721'
 projectId: u'apache-beam-testing'
 stageStates: []
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-05-25_08_59_17-15727091547768182389]
root: INFO: To access the Dataflow monitoring console, please navigate to https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-05-25_08_59_17-15727091547768182389
root: INFO: Job 2017-05-25_08_59_17-15727091547768182389 is in state JOB_STATE_RUNNING
root: INFO: 2017-05-25T15:59:17.847Z: JOB_MESSAGE_WARNING: (da41da6fdec55971): Setting the
number of workers (1) disables autoscaling for this job. If you are trying to cap autoscaling,
consider only setting max_num_workers. If you want to disable autoscaling altogether, the
documented way is to explicitly use autoscalingAlgorithm=NONE.
root: INFO: 2017-05-25T15:59:19.991Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0a26): Checking required
Cloud APIs are enabled.
root: INFO: 2017-05-25T15:59:21.027Z: JOB_MESSAGE_DEBUG: (b860003e1b5a0a51): Combiner lifting
skipped for step write/Write/WriteImpl/GroupByKey: GroupByKey not followed by a combiner.
root: INFO: 2017-05-25T15:59:21.029Z: JOB_MESSAGE_DEBUG: (b860003e1b5a0e9b): Combiner lifting
skipped for step group: GroupByKey not followed by a combiner.
root: INFO: 2017-05-25T15:59:21.031Z: JOB_MESSAGE_DETAILED: (b860003e1b5a02e5): Expanding
GroupByKey operations into optimizable parts.
root: INFO: 2017-05-25T15:59:21.034Z: JOB_MESSAGE_DETAILED: (b860003e1b5a072f): Lifting ValueCombiningMappingFns
into MergeBucketsMappingFns
root: INFO: 2017-05-25T15:59:21.041Z: JOB_MESSAGE_DEBUG: (b860003e1b5a040d): Annotating graph
with Autotuner information.
root: INFO: 2017-05-25T15:59:21.053Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0ca1): Fusing adjacent
ParDo, Read, Write, and Flatten operations
root: INFO: 2017-05-25T15:59:21.056Z: JOB_MESSAGE_DETAILED: (b860003e1b5a00eb): Fusing consumer
split into read/Read
root: INFO: 2017-05-25T15:59:21.058Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0535): Fusing consumer
group/Write into group/Reify
root: INFO: 2017-05-25T15:59:21.060Z: JOB_MESSAGE_DETAILED: (b860003e1b5a097f): Fusing consumer
group/GroupByWindow into group/Read
root: INFO: 2017-05-25T15:59:21.063Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0dc9): Fusing consumer
write/Write/WriteImpl/GroupByKey/GroupByWindow into write/Write/WriteImpl/GroupByKey/Read
root: INFO: 2017-05-25T15:59:21.066Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0213): Fusing consumer
write/Write/WriteImpl/GroupByKey/Write into write/Write/WriteImpl/GroupByKey/Reify
root: INFO: 2017-05-25T15:59:21.072Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0aa7): Fusing consumer
write/Write/WriteImpl/WindowInto(WindowIntoFn) into write/Write/WriteImpl/Pair
root: INFO: 2017-05-25T15:59:21.074Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0ef1): Fusing consumer
write/Write/WriteImpl/GroupByKey/Reify into write/Write/WriteImpl/WindowInto(WindowIntoFn)
root: INFO: 2017-05-25T15:59:21.077Z: JOB_MESSAGE_DETAILED: (b860003e1b5a033b): Fusing consumer
pair_with_one into split
root: INFO: 2017-05-25T15:59:21.079Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0785): Fusing consumer
group/Reify into pair_with_one
root: INFO: 2017-05-25T15:59:21.081Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0bcf): Fusing consumer
write/Write/WriteImpl/WriteBundles/Do into format
root: INFO: 2017-05-25T15:59:21.084Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0019): Fusing consumer
write/Write/WriteImpl/Pair into write/Write/WriteImpl/WriteBundles/Do
root: INFO: 2017-05-25T15:59:21.086Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0463): Fusing consumer
format into count
root: INFO: 2017-05-25T15:59:21.088Z: JOB_MESSAGE_DETAILED: (b860003e1b5a08ad): Fusing consumer
write/Write/WriteImpl/Extract into write/Write/WriteImpl/GroupByKey/GroupByWindow
root: INFO: 2017-05-25T15:59:21.091Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0cf7): Fusing consumer
count into group/GroupByWindow
root: INFO: 2017-05-25T15:59:21.100Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0e1f): Fusing consumer
write/Write/WriteImpl/InitializeWrite into write/Write/WriteImpl/DoOnce/Read
root: INFO: 2017-05-25T15:59:21.175Z: JOB_MESSAGE_DEBUG: (b860003e1b5a0ecb): Workflow config
is missing a default resource spec.
root: INFO: 2017-05-25T15:59:21.178Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0315): Adding StepResource
setup and teardown to workflow graph.
root: INFO: 2017-05-25T15:59:21.180Z: JOB_MESSAGE_DEBUG: (b860003e1b5a075f): Adding workflow
start and stop steps.
root: INFO: 2017-05-25T15:59:21.183Z: JOB_MESSAGE_DEBUG: (b860003e1b5a0ba9): Assigning stage
ids.
root: INFO: 2017-05-25T15:59:21.223Z: JOB_MESSAGE_DEBUG: (3c6c761f21815370): Executing wait
step start25
root: INFO: 2017-05-25T15:59:21.231Z: JOB_MESSAGE_BASIC: (3c6c761f2181586e): Executing operation
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
root: INFO: 2017-05-25T15:59:21.234Z: JOB_MESSAGE_BASIC: (741bbbb8c7e5509e): Executing operation
group/Create
root: INFO: 2017-05-25T15:59:21.435Z: JOB_MESSAGE_DEBUG: (5b2b3e4c17869251): Starting worker
pool setup.
root: INFO: 2017-05-25T15:59:21.437Z: JOB_MESSAGE_BASIC: (5b2b3e4c1786971f): Starting 1 workers...
root: INFO: 2017-05-25T15:59:21.449Z: JOB_MESSAGE_DEBUG: (741bbbb8c7e557cd): Value "group/Session"
materialized.
root: INFO: 2017-05-25T15:59:21.458Z: JOB_MESSAGE_BASIC: (741bbbb8c7e55385): Executing operation
read/Read+split+pair_with_one+group/Reify+group/Write
root: INFO: 2017-05-25T16:00:22.421Z: JOB_MESSAGE_DETAILED: (d1d7a890d50c72d6): Workers have
started successfully.
root: INFO: 2017-05-25T16:02:48.300Z: JOB_MESSAGE_ERROR: (7a3d72f03b4d4408): Traceback (most
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 706,
in run
    self._load_main_session(self.local_staging_directory)
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 446,
in _load_main_session
    pickler.load_session(session_file)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 247,
in load_session
    return dill.load_session(file_path)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 363, in load_session
    module = unpickler.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 766, in _import_module
    return __import__(import_name)
ImportError: No module named gen_protos

root: INFO: 2017-05-25T16:02:50.353Z: JOB_MESSAGE_ERROR: (7a3d72f03b4d4b40): Traceback (most
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 706,
in run
    self._load_main_session(self.local_staging_directory)
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 446,
in _load_main_session
    pickler.load_session(session_file)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 247,
in load_session
    return dill.load_session(file_path)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 363, in load_session
    module = unpickler.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 766, in _import_module
    return __import__(import_name)
ImportError: No module named gen_protos

root: INFO: 2017-05-25T16:02:52.407Z: JOB_MESSAGE_ERROR: (7a3d72f03b4d4278): Traceback (most
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 706,
in run
    self._load_main_session(self.local_staging_directory)
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 446,
in _load_main_session
    pickler.load_session(session_file)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 247,
in load_session
    return dill.load_session(file_path)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 363, in load_session
    module = unpickler.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 766, in _import_module
    return __import__(import_name)
ImportError: No module named gen_protos

root: INFO: 2017-05-25T16:02:54.460Z: JOB_MESSAGE_ERROR: (7a3d72f03b4d49b0): Traceback (most
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 706,
in run
    self._load_main_session(self.local_staging_directory)
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", line 446,
in _load_main_session
    pickler.load_session(session_file)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 247,
in load_session
    return dill.load_session(file_path)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 363, in load_session
    module = unpickler.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 766, in _import_module
    return __import__(import_name)
ImportError: No module named gen_protos

root: INFO: 2017-05-25T16:02:54.610Z: JOB_MESSAGE_DEBUG: (741bbbb8c7e551a2): Executing failure
step failure24
root: INFO: 2017-05-25T16:02:54.640Z: JOB_MESSAGE_ERROR: (741bbbb8c7e55890): Workflow failed.
Causes: (741bbbb8c7e553c6): S05:read/Read+split+pair_with_one+group/Reify+group/Write failed.,
(317ee513ba6bc569): Failed to split source.
root: INFO: 2017-05-25T16:02:54.721Z: JOB_MESSAGE_DETAILED: (b860003e1b5a0340): Cleaning up.
root: INFO: 2017-05-25T16:02:54.809Z: JOB_MESSAGE_DEBUG: (b860003e1b5a078a): Starting worker
pool teardown.
root: INFO: 2017-05-25T16:02:54.811Z: JOB_MESSAGE_BASIC: (b860003e1b5a0bd4): Stopping worker
pool...
root: INFO: 2017-05-25T16:04:09.828Z: JOB_MESSAGE_BASIC: (b860003e1b5a0243): Worker pool stopped.
root: INFO: 2017-05-25T16:04:09.866Z: JOB_MESSAGE_DEBUG: (b860003e1b5a036b): Tearing down
pending resources...
root: INFO: Job 2017-05-25_08_59_17-15727091547768182389 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 2 tests in 343.955s

FAILED (errors=1)
Found: https://console.cloud.google.com/dataflow/job/2017-05-25_08_59_16-12120469644462099244?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-05-25_08_59_17-15727091547768182389?project=apache-beam-testing
Build step 'Execute shell' marked build as failure

Mime
View raw message