beam-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Apache Jenkins Server <jenk...@builds.apache.org>
Subject Build failed in Jenkins: beam_PostCommit_Python_Verify #1935
Date Thu, 20 Apr 2017 16:55:14 GMT
See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/1935/display/redirect?page=changes>

Changes:

[robertwb] [BEAM-662] Fix for allowing floating point periods in windows

[robertwb] Make stage names consistent.

[robertwb] Require deterministic window coders.

[robertwb] Enable IntervalWindowCoder test check.

[robertwb] Remove obsolete and unused Runner.clear

[robertwb] Rename AfterFirst to AfterAny for consistency with Java.

[robertwb] Remove bigshuffle from python examples

[robertwb] Remove vestigial Read and Write from core.py

[aljoscha.krettek] [BEAM-1886] Remove TextIO override in Flink runner

[aljoscha.krettek] Exclude UsesSplittableParDoWithWindowedSideInputs in Flink Stream Runner

------------------------------------------
[...truncated 850.31 KB...]
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s12", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_merge_tagged_vals_under_key"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key).out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s11"
        }, 
        "serialized_fn": "<string of 1332 bytes>", 
        "user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key)"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s13", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Unkey.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s12"
        }, 
        "serialized_fn": "<string of 956 bytes>", 
        "user_name": "assert_that/Unkey"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s14", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_equal"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",

                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s13"
        }, 
        "serialized_fn": "<string of 1112 bytes>", 
        "user_name": "assert_that/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json
after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block',
'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin,
Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date':
'Thu, 20 Apr 2017 16:48:52 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json;
charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(3cd60379f74dbe03): The workflow could not be created. Causes: (66c8265657b52535):
Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for
active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow
to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json
after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '440', 'x-xss-protection': '1; mode=block',
'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin,
Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date':
'Thu, 20 Apr 2017 16:48:54 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json;
charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(3526da3a3bbd396): The workflow could not be created. Causes: (802261ba744c1a25):
Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for
active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow
to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json
after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block',
'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin,
Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date':
'Thu, 20 Apr 2017 16:49:03 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json;
charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(22c61baeeb4ac6db): The workflow could not be created. Causes: (71bb6e39d741719f):
Too many running jobs. Project apache-beam-testing is running 26 jobs and project limit for
active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow
to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json
after exception HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block',
'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin,
Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date':
'Thu, 20 Apr 2017 16:49:13 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json;
charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(639197c3a7b95033): The workflow could not be created. Causes: (31b3e2093d3627a8):
Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for
active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow
to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
root: ERROR: HTTP status 429 trying to create job at dataflow service endpoint https://dataflow.googleapis.com
root: CRITICAL: details of server error: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1; mode=block',
'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin,
Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date':
'Thu, 20 Apr 2017 16:49:32 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json;
charset=UTF-8'}>, content <{
  "error": {
    "code": 429,
    "message": "(e02f0d7f7d8f8e8e): The workflow could not be created. Causes: (5fb011768af83075):
Too many running jobs. Project apache-beam-testing is running 25 jobs and project limit for
active jobs is 25. To fix this, cancel an existing workflow via the UI, wait for a workflow
to finish or contact dataflow-feedback@google.com to request an increase in quota.",
    "status": "RESOURCE_EXHAUSTED"
  }
}
>
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 14 tests in 1143.787s

FAILED (errors=2)
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_36_03-17306565410535521102?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_42_25-11836023461597041757?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_48_50-13690095825366945233?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_36_03-12399630952851908062?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_42_45-4692674566775294509?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_36_03-11692027761086874489?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_42_20-4137566846231221883?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_47_02-15845890897779327383?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_36_03-8589621051685956297?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_41_46-12806371989050635570?project=apache-beam-testing
Found: https://console.cloud.google.com/dataflow/job/2017-04-20_09_46_58-9673877740759467903?project=apache-beam-testing
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user aljoscha.krettek@gmail.com

Mime
View raw message