beam-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From al...@apache.org
Subject [7/9] beam git commit: Move google specific io clients to vendor directory.
Date Sun, 19 Feb 2017 05:24:46 GMT
http://git-wip-us.apache.org/repos/asf/beam/blob/21925496/sdks/python/apache_beam/internal/clients/bigquery/bigquery_v2_messages.py
----------------------------------------------------------------------
diff --git a/sdks/python/apache_beam/internal/clients/bigquery/bigquery_v2_messages.py b/sdks/python/apache_beam/internal/clients/bigquery/bigquery_v2_messages.py
deleted file mode 100644
index 4045428..0000000
--- a/sdks/python/apache_beam/internal/clients/bigquery/bigquery_v2_messages.py
+++ /dev/null
@@ -1,1910 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#    http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-"""Generated message classes for bigquery version v2.
-
-A data platform for customers to create, manage, share and query data.
-"""
-# NOTE: This file is autogenerated and should not be edited by hand.
-
-from apitools.base.protorpclite import messages as _messages
-from apitools.base.py import encoding
-from apitools.base.py import extra_types
-
-
-package = 'bigquery'
-
-
-class BigqueryDatasetsDeleteRequest(_messages.Message):
-  """A BigqueryDatasetsDeleteRequest object.
-
-  Fields:
-    datasetId: Dataset ID of dataset being deleted
-    deleteContents: If True, delete all the tables in the dataset. If False
-      and the dataset contains tables, the request will fail. Default is False
-    projectId: Project ID of the dataset being deleted
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  deleteContents = _messages.BooleanField(2)
-  projectId = _messages.StringField(3, required=True)
-
-
-class BigqueryDatasetsDeleteResponse(_messages.Message):
-  """An empty BigqueryDatasetsDelete response."""
-
-
-class BigqueryDatasetsGetRequest(_messages.Message):
-  """A BigqueryDatasetsGetRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the requested dataset
-    projectId: Project ID of the requested dataset
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-
-
-class BigqueryDatasetsInsertRequest(_messages.Message):
-  """A BigqueryDatasetsInsertRequest object.
-
-  Fields:
-    dataset: A Dataset resource to be passed as the request body.
-    projectId: Project ID of the new dataset
-  """
-
-  dataset = _messages.MessageField('Dataset', 1)
-  projectId = _messages.StringField(2, required=True)
-
-
-class BigqueryDatasetsListRequest(_messages.Message):
-  """A BigqueryDatasetsListRequest object.
-
-  Fields:
-    all: Whether to list all datasets, including hidden ones
-    maxResults: The maximum number of results to return
-    pageToken: Page token, returned by a previous call, to request the next
-      page of results
-    projectId: Project ID of the datasets to be listed
-  """
-
-  all = _messages.BooleanField(1)
-  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
-  pageToken = _messages.StringField(3)
-  projectId = _messages.StringField(4, required=True)
-
-
-class BigqueryDatasetsPatchRequest(_messages.Message):
-  """A BigqueryDatasetsPatchRequest object.
-
-  Fields:
-    dataset: A Dataset resource to be passed as the request body.
-    datasetId: Dataset ID of the dataset being updated
-    projectId: Project ID of the dataset being updated
-  """
-
-  dataset = _messages.MessageField('Dataset', 1)
-  datasetId = _messages.StringField(2, required=True)
-  projectId = _messages.StringField(3, required=True)
-
-
-class BigqueryDatasetsUpdateRequest(_messages.Message):
-  """A BigqueryDatasetsUpdateRequest object.
-
-  Fields:
-    dataset: A Dataset resource to be passed as the request body.
-    datasetId: Dataset ID of the dataset being updated
-    projectId: Project ID of the dataset being updated
-  """
-
-  dataset = _messages.MessageField('Dataset', 1)
-  datasetId = _messages.StringField(2, required=True)
-  projectId = _messages.StringField(3, required=True)
-
-
-class BigqueryJobsCancelRequest(_messages.Message):
-  """A BigqueryJobsCancelRequest object.
-
-  Fields:
-    jobId: [Required] Job ID of the job to cancel
-    projectId: [Required] Project ID of the job to cancel
-  """
-
-  jobId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-
-
-class BigqueryJobsGetQueryResultsRequest(_messages.Message):
-  """A BigqueryJobsGetQueryResultsRequest object.
-
-  Fields:
-    jobId: [Required] Job ID of the query job
-    maxResults: Maximum number of results to read
-    pageToken: Page token, returned by a previous call, to request the next
-      page of results
-    projectId: [Required] Project ID of the query job
-    startIndex: Zero-based index of the starting row
-    timeoutMs: How long to wait for the query to complete, in milliseconds,
-      before returning. Default is 10 seconds. If the timeout passes before
-      the job completes, the 'jobComplete' field in the response will be false
-  """
-
-  jobId = _messages.StringField(1, required=True)
-  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
-  pageToken = _messages.StringField(3)
-  projectId = _messages.StringField(4, required=True)
-  startIndex = _messages.IntegerField(5, variant=_messages.Variant.UINT64)
-  timeoutMs = _messages.IntegerField(6, variant=_messages.Variant.UINT32)
-
-
-class BigqueryJobsGetRequest(_messages.Message):
-  """A BigqueryJobsGetRequest object.
-
-  Fields:
-    jobId: [Required] Job ID of the requested job
-    projectId: [Required] Project ID of the requested job
-  """
-
-  jobId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-
-
-class BigqueryJobsInsertRequest(_messages.Message):
-  """A BigqueryJobsInsertRequest object.
-
-  Fields:
-    job: A Job resource to be passed as the request body.
-    projectId: Project ID of the project that will be billed for the job
-  """
-
-  job = _messages.MessageField('Job', 1)
-  projectId = _messages.StringField(2, required=True)
-
-
-class BigqueryJobsListRequest(_messages.Message):
-  """A BigqueryJobsListRequest object.
-
-  Enums:
-    ProjectionValueValuesEnum: Restrict information returned to a set of
-      selected fields
-    StateFilterValueValuesEnum: Filter for job state
-
-  Fields:
-    allUsers: Whether to display jobs owned by all users in the project.
-      Default false
-    maxResults: Maximum number of results to return
-    pageToken: Page token, returned by a previous call, to request the next
-      page of results
-    projectId: Project ID of the jobs to list
-    projection: Restrict information returned to a set of selected fields
-    stateFilter: Filter for job state
-  """
-
-  class ProjectionValueValuesEnum(_messages.Enum):
-    """Restrict information returned to a set of selected fields
-
-    Values:
-      full: Includes all job data
-      minimal: Does not include the job configuration
-    """
-    full = 0
-    minimal = 1
-
-  class StateFilterValueValuesEnum(_messages.Enum):
-    """Filter for job state
-
-    Values:
-      done: Finished jobs
-      pending: Pending jobs
-      running: Running jobs
-    """
-    done = 0
-    pending = 1
-    running = 2
-
-  allUsers = _messages.BooleanField(1)
-  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
-  pageToken = _messages.StringField(3)
-  projectId = _messages.StringField(4, required=True)
-  projection = _messages.EnumField('ProjectionValueValuesEnum', 5)
-  stateFilter = _messages.EnumField('StateFilterValueValuesEnum', 6, repeated=True)
-
-
-class BigqueryJobsQueryRequest(_messages.Message):
-  """A BigqueryJobsQueryRequest object.
-
-  Fields:
-    projectId: Project ID of the project billed for the query
-    queryRequest: A QueryRequest resource to be passed as the request body.
-  """
-
-  projectId = _messages.StringField(1, required=True)
-  queryRequest = _messages.MessageField('QueryRequest', 2)
-
-
-class BigqueryProjectsListRequest(_messages.Message):
-  """A BigqueryProjectsListRequest object.
-
-  Fields:
-    maxResults: Maximum number of results to return
-    pageToken: Page token, returned by a previous call, to request the next
-      page of results
-  """
-
-  maxResults = _messages.IntegerField(1, variant=_messages.Variant.UINT32)
-  pageToken = _messages.StringField(2)
-
-
-class BigqueryTabledataInsertAllRequest(_messages.Message):
-  """A BigqueryTabledataInsertAllRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the destination table.
-    projectId: Project ID of the destination table.
-    tableDataInsertAllRequest: A TableDataInsertAllRequest resource to be
-      passed as the request body.
-    tableId: Table ID of the destination table.
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-  tableDataInsertAllRequest = _messages.MessageField('TableDataInsertAllRequest', 3)
-  tableId = _messages.StringField(4, required=True)
-
-
-class BigqueryTabledataListRequest(_messages.Message):
-  """A BigqueryTabledataListRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the table to read
-    maxResults: Maximum number of results to return
-    pageToken: Page token, returned by a previous call, identifying the result
-      set
-    projectId: Project ID of the table to read
-    startIndex: Zero-based index of the starting row to read
-    tableId: Table ID of the table to read
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
-  pageToken = _messages.StringField(3)
-  projectId = _messages.StringField(4, required=True)
-  startIndex = _messages.IntegerField(5, variant=_messages.Variant.UINT64)
-  tableId = _messages.StringField(6, required=True)
-
-
-class BigqueryTablesDeleteRequest(_messages.Message):
-  """A BigqueryTablesDeleteRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the table to delete
-    projectId: Project ID of the table to delete
-    tableId: Table ID of the table to delete
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-  tableId = _messages.StringField(3, required=True)
-
-
-class BigqueryTablesDeleteResponse(_messages.Message):
-  """An empty BigqueryTablesDelete response."""
-
-
-class BigqueryTablesGetRequest(_messages.Message):
-  """A BigqueryTablesGetRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the requested table
-    projectId: Project ID of the requested table
-    tableId: Table ID of the requested table
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-  tableId = _messages.StringField(3, required=True)
-
-
-class BigqueryTablesInsertRequest(_messages.Message):
-  """A BigqueryTablesInsertRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the new table
-    projectId: Project ID of the new table
-    table: A Table resource to be passed as the request body.
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-  table = _messages.MessageField('Table', 3)
-
-
-class BigqueryTablesListRequest(_messages.Message):
-  """A BigqueryTablesListRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the tables to list
-    maxResults: Maximum number of results to return
-    pageToken: Page token, returned by a previous call, to request the next
-      page of results
-    projectId: Project ID of the tables to list
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
-  pageToken = _messages.StringField(3)
-  projectId = _messages.StringField(4, required=True)
-
-
-class BigqueryTablesPatchRequest(_messages.Message):
-  """A BigqueryTablesPatchRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the table to update
-    projectId: Project ID of the table to update
-    table: A Table resource to be passed as the request body.
-    tableId: Table ID of the table to update
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-  table = _messages.MessageField('Table', 3)
-  tableId = _messages.StringField(4, required=True)
-
-
-class BigqueryTablesUpdateRequest(_messages.Message):
-  """A BigqueryTablesUpdateRequest object.
-
-  Fields:
-    datasetId: Dataset ID of the table to update
-    projectId: Project ID of the table to update
-    table: A Table resource to be passed as the request body.
-    tableId: Table ID of the table to update
-  """
-
-  datasetId = _messages.StringField(1, required=True)
-  projectId = _messages.StringField(2, required=True)
-  table = _messages.MessageField('Table', 3)
-  tableId = _messages.StringField(4, required=True)
-
-
-class BigtableColumn(_messages.Message):
-  """A BigtableColumn object.
-
-  Fields:
-    encoding: [Optional] The encoding of the values when the type is not
-      STRING. Acceptable encoding values are: TEXT - indicates values are
-      alphanumeric text strings. BINARY - indicates values are encoded using
-      HBase Bytes.toBytes family of functions. 'encoding' can also be set at
-      the column family level. However, the setting at this level takes
-      precedence if 'encoding' is set at both levels.
-    fieldName: [Optional] If the qualifier is not a valid BigQuery field
-      identifier i.e. does not match [a-zA-Z][a-zA-Z0-9_]*, a valid identifier
-      must be provided as the column field name and is used as field name in
-      queries.
-    onlyReadLatest: [Optional] If this is set, only the latest version of
-      value in this column are exposed. 'onlyReadLatest' can also be set at
-      the column family level. However, the setting at this level takes
-      precedence if 'onlyReadLatest' is set at both levels.
-    qualifierEncoded: [Required] Qualifier of the column. Columns in the
-      parent column family that has this exact qualifier are exposed as .
-      field. If the qualifier is valid UTF-8 string, it can be specified in
-      the qualifier_string field. Otherwise, a base-64 encoded value must be
-      set to qualifier_encoded. The column field name is the same as the
-      column qualifier. However, if the qualifier is not a valid BigQuery
-      field identifier i.e. does not match [a-zA-Z][a-zA-Z0-9_]*, a valid
-      identifier must be provided as field_name.
-    qualifierString: A string attribute.
-    type: [Optional] The type to convert the value in cells of this column.
-      The values are expected to be encoded using HBase Bytes.toBytes function
-      when using the BINARY encoding value. Following BigQuery types are
-      allowed (case-sensitive) - BYTES STRING INTEGER FLOAT BOOLEAN Defaut
-      type is BYTES. 'type' can also be set at the column family level.
-      However, the setting at this level takes precedence if 'type' is set at
-      both levels.
-  """
-
-  encoding = _messages.StringField(1)
-  fieldName = _messages.StringField(2)
-  onlyReadLatest = _messages.BooleanField(3)
-  qualifierEncoded = _messages.BytesField(4)
-  qualifierString = _messages.StringField(5)
-  type = _messages.StringField(6)
-
-
-class BigtableColumnFamily(_messages.Message):
-  """A BigtableColumnFamily object.
-
-  Fields:
-    columns: [Optional] Lists of columns that should be exposed as individual
-      fields as opposed to a list of (column name, value) pairs. All columns
-      whose qualifier matches a qualifier in this list can be accessed as ..
-      Other columns can be accessed as a list through .Column field.
-    encoding: [Optional] The encoding of the values when the type is not
-      STRING. Acceptable encoding values are: TEXT - indicates values are
-      alphanumeric text strings. BINARY - indicates values are encoded using
-      HBase Bytes.toBytes family of functions. This can be overridden for a
-      specific column by listing that column in 'columns' and specifying an
-      encoding for it.
-    familyId: Identifier of the column family.
-    onlyReadLatest: [Optional] If this is set only the latest version of value
-      are exposed for all columns in this column family. This can be
-      overridden for a specific column by listing that column in 'columns' and
-      specifying a different setting for that column.
-    type: [Optional] The type to convert the value in cells of this column
-      family. The values are expected to be encoded using HBase Bytes.toBytes
-      function when using the BINARY encoding value. Following BigQuery types
-      are allowed (case-sensitive) - BYTES STRING INTEGER FLOAT BOOLEAN Defaut
-      type is BYTES. This can be overridden for a specific column by listing
-      that column in 'columns' and specifying a type for it.
-  """
-
-  columns = _messages.MessageField('BigtableColumn', 1, repeated=True)
-  encoding = _messages.StringField(2)
-  familyId = _messages.StringField(3)
-  onlyReadLatest = _messages.BooleanField(4)
-  type = _messages.StringField(5)
-
-
-class BigtableOptions(_messages.Message):
-  """A BigtableOptions object.
-
-  Fields:
-    columnFamilies: [Optional] List of column families to expose in the table
-      schema along with their types. This list restricts the column families
-      that can be referenced in queries and specifies their value types. You
-      can use this list to do type conversions - see the 'type' field for more
-      details. If you leave this list empty, all column families are present
-      in the table schema and their values are read as BYTES. During a query
-      only the column families referenced in that query are read from
-      Bigtable.
-    ignoreUnspecifiedColumnFamilies: [Optional] If field is true, then the
-      column families that are not specified in columnFamilies list are not
-      exposed in the table schema. Otherwise, they are read with BYTES type
-      values. The default value is false.
-  """
-
-  columnFamilies = _messages.MessageField('BigtableColumnFamily', 1, repeated=True)
-  ignoreUnspecifiedColumnFamilies = _messages.BooleanField(2)
-
-
-class CsvOptions(_messages.Message):
-  """A CsvOptions object.
-
-  Fields:
-    allowJaggedRows: [Optional] Indicates if BigQuery should accept rows that
-      are missing trailing optional columns. If true, BigQuery treats missing
-      trailing columns as null values. If false, records with missing trailing
-      columns are treated as bad records, and if there are too many bad
-      records, an invalid error is returned in the job result. The default
-      value is false.
-    allowQuotedNewlines: [Optional] Indicates if BigQuery should allow quoted
-      data sections that contain newline characters in a CSV file. The default
-      value is false.
-    encoding: [Optional] The character encoding of the data. The supported
-      values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery
-      decodes the data after the raw, binary data has been split using the
-      values of the quote and fieldDelimiter properties.
-    fieldDelimiter: [Optional] The separator for fields in a CSV file.
-      BigQuery converts the string to ISO-8859-1 encoding, and then uses the
-      first byte of the encoded string to split the data in its raw, binary
-      state. BigQuery also supports the escape sequence "\t" to specify a tab
-      separator. The default value is a comma (',').
-    quote: [Optional] The value that is used to quote data sections in a CSV
-      file. BigQuery converts the string to ISO-8859-1 encoding, and then uses
-      the first byte of the encoded string to split the data in its raw,
-      binary state. The default value is a double-quote ('"'). If your data
-      does not contain quoted sections, set the property value to an empty
-      string. If your data contains quoted newline characters, you must also
-      set the allowQuotedNewlines property to true.
-    skipLeadingRows: [Optional] The number of rows at the top of a CSV file
-      that BigQuery will skip when reading the data. The default value is 0.
-      This property is useful if you have header rows in the file that should
-      be skipped.
-  """
-
-  allowJaggedRows = _messages.BooleanField(1)
-  allowQuotedNewlines = _messages.BooleanField(2)
-  encoding = _messages.StringField(3)
-  fieldDelimiter = _messages.StringField(4)
-  quote = _messages.StringField(5, default=u'"')
-  skipLeadingRows = _messages.IntegerField(6, variant=_messages.Variant.INT32)
-
-
-class Dataset(_messages.Message):
-  """A Dataset object.
-
-  Messages:
-    AccessValueListEntry: A AccessValueListEntry object.
-
-  Fields:
-    access: [Optional] An array of objects that define dataset access for one
-      or more entities. You can set this property when inserting or updating a
-      dataset in order to control who is allowed to access the data. If
-      unspecified at dataset creation time, BigQuery adds default dataset
-      access for the following entities: access.specialGroup: projectReaders;
-      access.role: READER; access.specialGroup: projectWriters; access.role:
-      WRITER; access.specialGroup: projectOwners; access.role: OWNER;
-      access.userByEmail: [dataset creator email]; access.role: OWNER;
-    creationTime: [Output-only] The time when this dataset was created, in
-      milliseconds since the epoch.
-    datasetReference: [Required] A reference that identifies the dataset.
-    defaultTableExpirationMs: [Optional] The default lifetime of all tables in
-      the dataset, in milliseconds. The minimum value is 3600000 milliseconds
-      (one hour). Once this property is set, all newly-created tables in the
-      dataset will have an expirationTime property set to the creation time
-      plus the value in this property, and changing the value will only affect
-      new tables, not existing ones. When the expirationTime for a given table
-      is reached, that table will be deleted automatically. If a table's
-      expirationTime is modified or removed before the table expires, or if
-      you provide an explicit expirationTime when creating a table, that value
-      takes precedence over the default expiration time indicated by this
-      property.
-    description: [Optional] A user-friendly description of the dataset.
-    etag: [Output-only] A hash of the resource.
-    friendlyName: [Optional] A descriptive name for the dataset.
-    id: [Output-only] The fully-qualified unique name of the dataset in the
-      format projectId:datasetId. The dataset name without the project name is
-      given in the datasetId field. When creating a new dataset, leave this
-      field blank, and instead specify the datasetId field.
-    kind: [Output-only] The resource type.
-    lastModifiedTime: [Output-only] The date when this dataset or any of its
-      tables was last modified, in milliseconds since the epoch.
-    location: [Experimental] The geographic location where the dataset should
-      reside. Possible values include EU and US. The default value is US.
-    selfLink: [Output-only] A URL that can be used to access the resource
-      again. You can use this URL in Get or Update requests to the resource.
-  """
-
-  class AccessValueListEntry(_messages.Message):
-    """A AccessValueListEntry object.
-
-    Fields:
-      domain: [Pick one] A domain to grant access to. Any users signed in with
-        the domain specified will be granted the specified access. Example:
-        "example.com".
-      groupByEmail: [Pick one] An email address of a Google Group to grant
-        access to.
-      role: [Required] Describes the rights granted to the user specified by
-        the other member of the access object. The following string values are
-        supported: READER, WRITER, OWNER.
-      specialGroup: [Pick one] A special group to grant access to. Possible
-        values include: projectOwners: Owners of the enclosing project.
-        projectReaders: Readers of the enclosing project. projectWriters:
-        Writers of the enclosing project. allAuthenticatedUsers: All
-        authenticated BigQuery users.
-      userByEmail: [Pick one] An email address of a user to grant access to.
-        For example: fred@example.com.
-      view: [Pick one] A view from a different dataset to grant access to.
-        Queries executed against that view will have read access to tables in
-        this dataset. The role field is not required when this field is set.
-        If that view is updated by any user, access to the view needs to be
-        granted again via an update operation.
-    """
-
-    domain = _messages.StringField(1)
-    groupByEmail = _messages.StringField(2)
-    role = _messages.StringField(3)
-    specialGroup = _messages.StringField(4)
-    userByEmail = _messages.StringField(5)
-    view = _messages.MessageField('TableReference', 6)
-
-  access = _messages.MessageField('AccessValueListEntry', 1, repeated=True)
-  creationTime = _messages.IntegerField(2)
-  datasetReference = _messages.MessageField('DatasetReference', 3)
-  defaultTableExpirationMs = _messages.IntegerField(4)
-  description = _messages.StringField(5)
-  etag = _messages.StringField(6)
-  friendlyName = _messages.StringField(7)
-  id = _messages.StringField(8)
-  kind = _messages.StringField(9, default=u'bigquery#dataset')
-  lastModifiedTime = _messages.IntegerField(10)
-  location = _messages.StringField(11)
-  selfLink = _messages.StringField(12)
-
-
-class DatasetList(_messages.Message):
-  """A DatasetList object.
-
-  Messages:
-    DatasetsValueListEntry: A DatasetsValueListEntry object.
-
-  Fields:
-    datasets: An array of the dataset resources in the project. Each resource
-      contains basic information. For full information about a particular
-      dataset resource, use the Datasets: get method. This property is omitted
-      when there are no datasets in the project.
-    etag: A hash value of the results page. You can use this property to
-      determine if the page has changed since the last request.
-    kind: The list type. This property always returns the value
-      "bigquery#datasetList".
-    nextPageToken: A token that can be used to request the next results page.
-      This property is omitted on the final results page.
-  """
-
-  class DatasetsValueListEntry(_messages.Message):
-    """A DatasetsValueListEntry object.
-
-    Fields:
-      datasetReference: The dataset reference. Use this property to access
-        specific parts of the dataset's ID, such as project ID or dataset ID.
-      friendlyName: A descriptive name for the dataset, if one exists.
-      id: The fully-qualified, unique, opaque ID of the dataset.
-      kind: The resource type. This property always returns the value
-        "bigquery#dataset".
-    """
-
-    datasetReference = _messages.MessageField('DatasetReference', 1)
-    friendlyName = _messages.StringField(2)
-    id = _messages.StringField(3)
-    kind = _messages.StringField(4, default=u'bigquery#dataset')
-
-  datasets = _messages.MessageField('DatasetsValueListEntry', 1, repeated=True)
-  etag = _messages.StringField(2)
-  kind = _messages.StringField(3, default=u'bigquery#datasetList')
-  nextPageToken = _messages.StringField(4)
-
-
-class DatasetReference(_messages.Message):
-  """A DatasetReference object.
-
-  Fields:
-    datasetId: [Required] A unique ID for this dataset, without the project
-      name. The ID must contain only letters (a-z, A-Z), numbers (0-9), or
-      underscores (_). The maximum length is 1,024 characters.
-    projectId: [Optional] The ID of the project containing this dataset.
-  """
-
-  datasetId = _messages.StringField(1)
-  projectId = _messages.StringField(2)
-
-
-class ErrorProto(_messages.Message):
-  """A ErrorProto object.
-
-  Fields:
-    debugInfo: Debugging information. This property is internal to Google and
-      should not be used.
-    location: Specifies where the error occurred, if present.
-    message: A human-readable description of the error.
-    reason: A short error code that summarizes the error.
-  """
-
-  debugInfo = _messages.StringField(1)
-  location = _messages.StringField(2)
-  message = _messages.StringField(3)
-  reason = _messages.StringField(4)
-
-
-class ExplainQueryStage(_messages.Message):
-  """A ExplainQueryStage object.
-
-  Fields:
-    computeRatioAvg: Relative amount of time the average shard spent on CPU-
-      bound tasks.
-    computeRatioMax: Relative amount of time the slowest shard spent on CPU-
-      bound tasks.
-    id: Unique ID for stage within plan.
-    name: Human-readable name for stage.
-    readRatioAvg: Relative amount of time the average shard spent reading
-      input.
-    readRatioMax: Relative amount of time the slowest shard spent reading
-      input.
-    recordsRead: Number of records read into the stage.
-    recordsWritten: Number of records written by the stage.
-    steps: List of operations within the stage in dependency order
-      (approximately chronological).
-    waitRatioAvg: Relative amount of time the average shard spent waiting to
-      be scheduled.
-    waitRatioMax: Relative amount of time the slowest shard spent waiting to
-      be scheduled.
-    writeRatioAvg: Relative amount of time the average shard spent on writing
-      output.
-    writeRatioMax: Relative amount of time the slowest shard spent on writing
-      output.
-  """
-
-  computeRatioAvg = _messages.FloatField(1)
-  computeRatioMax = _messages.FloatField(2)
-  id = _messages.IntegerField(3)
-  name = _messages.StringField(4)
-  readRatioAvg = _messages.FloatField(5)
-  readRatioMax = _messages.FloatField(6)
-  recordsRead = _messages.IntegerField(7)
-  recordsWritten = _messages.IntegerField(8)
-  steps = _messages.MessageField('ExplainQueryStep', 9, repeated=True)
-  waitRatioAvg = _messages.FloatField(10)
-  waitRatioMax = _messages.FloatField(11)
-  writeRatioAvg = _messages.FloatField(12)
-  writeRatioMax = _messages.FloatField(13)
-
-
-class ExplainQueryStep(_messages.Message):
-  """A ExplainQueryStep object.
-
-  Fields:
-    kind: Machine-readable operation type.
-    substeps: Human-readable stage descriptions.
-  """
-
-  kind = _messages.StringField(1)
-  substeps = _messages.StringField(2, repeated=True)
-
-
-class ExternalDataConfiguration(_messages.Message):
-  """A ExternalDataConfiguration object.
-
-  Fields:
-    autodetect: [Experimental] Try to detect schema and format options
-      automatically. Any option specified explicitly will be honored.
-    bigtableOptions: [Optional] Additional options if sourceFormat is set to
-      BIGTABLE.
-    compression: [Optional] The compression type of the data source. Possible
-      values include GZIP and NONE. The default value is NONE. This setting is
-      ignored for Google Cloud Bigtable, Google Cloud Datastore backups and
-      Avro formats.
-    csvOptions: Additional properties to set if sourceFormat is set to CSV.
-    ignoreUnknownValues: [Optional] Indicates if BigQuery should allow extra
-      values that are not represented in the table schema. If true, the extra
-      values are ignored. If false, records with extra columns are treated as
-      bad records, and if there are too many bad records, an invalid error is
-      returned in the job result. The default value is false. The sourceFormat
-      property determines what BigQuery treats as an extra value: CSV:
-      Trailing columns JSON: Named values that don't match any column names
-      Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore
-      backups: This setting is ignored. Avro: This setting is ignored.
-    maxBadRecords: [Optional] The maximum number of bad records that BigQuery
-      can ignore when reading data. If the number of bad records exceeds this
-      value, an invalid error is returned in the job result. The default value
-      is 0, which requires that all records are valid. This setting is ignored
-      for Google Cloud Bigtable, Google Cloud Datastore backups and Avro
-      formats.
-    schema: [Optional] The schema for the data. Schema is required for CSV and
-      JSON formats. Schema is disallowed for Google Cloud Bigtable, Cloud
-      Datastore backups, and Avro formats.
-    sourceFormat: [Required] The data format. For CSV files, specify "CSV".
-      For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro
-      files, specify "AVRO". For Google Cloud Datastore backups, specify
-      "DATASTORE_BACKUP". [Experimental] For Google Cloud Bigtable, specify
-      "BIGTABLE". Please note that reading from Google Cloud Bigtable is
-      experimental and has to be enabled for your project. Please contact
-      Google Cloud Support to enable this for your project.
-    sourceUris: [Required] The fully-qualified URIs that point to your data in
-      Google Cloud. For Google Cloud Storage URIs: Each URI can contain one
-      '*' wildcard character and it must come after the 'bucket' name. Size
-      limits related to load jobs apply to external data sources, plus an
-      additional limit of 10 GB maximum size across all URIs. For Google Cloud
-      Bigtable URIs: Exactly one URI can be specified and it has be a fully
-      specified and valid HTTPS URL for a Google Cloud Bigtable table. For
-      Google Cloud Datastore backups, exactly one URI can be specified, and it
-      must end with '.backup_info'. Also, the '*' wildcard character is not
-      allowed.
-  """
-
-  autodetect = _messages.BooleanField(1)
-  bigtableOptions = _messages.MessageField('BigtableOptions', 2)
-  compression = _messages.StringField(3)
-  csvOptions = _messages.MessageField('CsvOptions', 4)
-  ignoreUnknownValues = _messages.BooleanField(5)
-  maxBadRecords = _messages.IntegerField(6, variant=_messages.Variant.INT32)
-  schema = _messages.MessageField('TableSchema', 7)
-  sourceFormat = _messages.StringField(8)
-  sourceUris = _messages.StringField(9, repeated=True)
-
-
-class GetQueryResultsResponse(_messages.Message):
-  """A GetQueryResultsResponse object.
-
-  Fields:
-    cacheHit: Whether the query result was fetched from the query cache.
-    errors: [Output-only] All errors and warnings encountered during the
-      running of the job. Errors here do not necessarily mean that the job has
-      completed or was unsuccessful.
-    etag: A hash of this response.
-    jobComplete: Whether the query has completed or not. If rows or totalRows
-      are present, this will always be true. If this is false, totalRows will
-      not be available.
-    jobReference: Reference to the BigQuery Job that was created to run the
-      query. This field will be present even if the original request timed
-      out, in which case GetQueryResults can be used to read the results once
-      the query has completed. Since this API only returns the first page of
-      results, subsequent pages can be fetched via the same mechanism
-      (GetQueryResults).
-    kind: The resource type of the response.
-    pageToken: A token used for paging results.
-    rows: An object with as many results as can be contained within the
-      maximum permitted reply size. To get any additional rows, you can call
-      GetQueryResults and specify the jobReference returned above. Present
-      only when the query completes successfully.
-    schema: The schema of the results. Present only when the query completes
-      successfully.
-    totalBytesProcessed: The total number of bytes processed for this query.
-    totalRows: The total number of rows in the complete query result set,
-      which can be more than the number of rows in this single page of
-      results. Present only when the query completes successfully.
-  """
-
-  cacheHit = _messages.BooleanField(1)
-  errors = _messages.MessageField('ErrorProto', 2, repeated=True)
-  etag = _messages.StringField(3)
-  jobComplete = _messages.BooleanField(4)
-  jobReference = _messages.MessageField('JobReference', 5)
-  kind = _messages.StringField(6, default=u'bigquery#getQueryResultsResponse')
-  pageToken = _messages.StringField(7)
-  rows = _messages.MessageField('TableRow', 8, repeated=True)
-  schema = _messages.MessageField('TableSchema', 9)
-  totalBytesProcessed = _messages.IntegerField(10)
-  totalRows = _messages.IntegerField(11, variant=_messages.Variant.UINT64)
-
-
-class IntervalPartitionConfiguration(_messages.Message):
-  """A IntervalPartitionConfiguration object.
-
-  Fields:
-    expirationMs: A string attribute.
-    type: A string attribute.
-  """
-
-  expirationMs = _messages.IntegerField(1)
-  type = _messages.StringField(2)
-
-
-class Job(_messages.Message):
-  """A Job object.
-
-  Fields:
-    configuration: [Required] Describes the job configuration.
-    etag: [Output-only] A hash of this resource.
-    id: [Output-only] Opaque ID field of the job
-    jobReference: [Optional] Reference describing the unique-per-user name of
-      the job.
-    kind: [Output-only] The type of the resource.
-    selfLink: [Output-only] A URL that can be used to access this resource
-      again.
-    statistics: [Output-only] Information about the job, including starting
-      time and ending time of the job.
-    status: [Output-only] The status of this job. Examine this value when
-      polling an asynchronous job to see if the job is complete.
-    user_email: [Output-only] Email address of the user who ran the job.
-  """
-
-  configuration = _messages.MessageField('JobConfiguration', 1)
-  etag = _messages.StringField(2)
-  id = _messages.StringField(3)
-  jobReference = _messages.MessageField('JobReference', 4)
-  kind = _messages.StringField(5, default=u'bigquery#job')
-  selfLink = _messages.StringField(6)
-  statistics = _messages.MessageField('JobStatistics', 7)
-  status = _messages.MessageField('JobStatus', 8)
-  user_email = _messages.StringField(9)
-
-
-class JobCancelResponse(_messages.Message):
-  """A JobCancelResponse object.
-
-  Fields:
-    job: The final state of the job.
-    kind: The resource type of the response.
-  """
-
-  job = _messages.MessageField('Job', 1)
-  kind = _messages.StringField(2, default=u'bigquery#jobCancelResponse')
-
-
-class JobConfiguration(_messages.Message):
-  """A JobConfiguration object.
-
-  Fields:
-    copy: [Pick one] Copies a table.
-    dryRun: [Optional] If set, don't actually run this job. A valid query will
-      return a mostly empty response with some processing statistics, while an
-      invalid query will return the same error it would if it wasn't a dry
-      run. Behavior of non-query jobs is undefined.
-    extract: [Pick one] Configures an extract job.
-    load: [Pick one] Configures a load job.
-    query: [Pick one] Configures a query job.
-  """
-
-  copy = _messages.MessageField('JobConfigurationTableCopy', 1)
-  dryRun = _messages.BooleanField(2)
-  extract = _messages.MessageField('JobConfigurationExtract', 3)
-  load = _messages.MessageField('JobConfigurationLoad', 4)
-  query = _messages.MessageField('JobConfigurationQuery', 5)
-
-
-class JobConfigurationExtract(_messages.Message):
-  """A JobConfigurationExtract object.
-
-  Fields:
-    compression: [Optional] The compression type to use for exported files.
-      Possible values include GZIP and NONE. The default value is NONE.
-    destinationFormat: [Optional] The exported file format. Possible values
-      include CSV, NEWLINE_DELIMITED_JSON and AVRO. The default value is CSV.
-      Tables with nested or repeated fields cannot be exported as CSV.
-    destinationUri: [Pick one] DEPRECATED: Use destinationUris instead,
-      passing only one URI as necessary. The fully-qualified Google Cloud
-      Storage URI where the extracted table should be written.
-    destinationUris: [Pick one] A list of fully-qualified Google Cloud Storage
-      URIs where the extracted table should be written.
-    fieldDelimiter: [Optional] Delimiter to use between fields in the exported
-      data. Default is ','
-    printHeader: [Optional] Whether to print out a header row in the results.
-      Default is true.
-    sourceTable: [Required] A reference to the table being exported.
-  """
-
-  compression = _messages.StringField(1)
-  destinationFormat = _messages.StringField(2)
-  destinationUri = _messages.StringField(3)
-  destinationUris = _messages.StringField(4, repeated=True)
-  fieldDelimiter = _messages.StringField(5)
-  printHeader = _messages.BooleanField(6, default=True)
-  sourceTable = _messages.MessageField('TableReference', 7)
-
-
-class JobConfigurationLoad(_messages.Message):
-  """A JobConfigurationLoad object.
-
-  Fields:
-    allowJaggedRows: [Optional] Accept rows that are missing trailing optional
-      columns. The missing values are treated as nulls. If false, records with
-      missing trailing columns are treated as bad records, and if there are
-      too many bad records, an invalid error is returned in the job result.
-      The default value is false. Only applicable to CSV, ignored for other
-      formats.
-    allowQuotedNewlines: Indicates if BigQuery should allow quoted data
-      sections that contain newline characters in a CSV file. The default
-      value is false.
-    createDisposition: [Optional] Specifies whether the job is allowed to
-      create new tables. The following values are supported: CREATE_IF_NEEDED:
-      If the table does not exist, BigQuery creates the table. CREATE_NEVER:
-      The table must already exist. If it does not, a 'notFound' error is
-      returned in the job result. The default value is CREATE_IF_NEEDED.
-      Creation, truncation and append actions occur as one atomic update upon
-      job completion.
-    destinationTable: [Required] The destination table to load the data into.
-    encoding: [Optional] The character encoding of the data. The supported
-      values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery
-      decodes the data after the raw, binary data has been split using the
-      values of the quote and fieldDelimiter properties.
-    fieldDelimiter: [Optional] The separator for fields in a CSV file. The
-      separator can be any ISO-8859-1 single-byte character. To use a
-      character in the range 128-255, you must encode the character as UTF8.
-      BigQuery converts the string to ISO-8859-1 encoding, and then uses the
-      first byte of the encoded string to split the data in its raw, binary
-      state. BigQuery also supports the escape sequence "\t" to specify a tab
-      separator. The default value is a comma (',').
-    ignoreUnknownValues: [Optional] Indicates if BigQuery should allow extra
-      values that are not represented in the table schema. If true, the extra
-      values are ignored. If false, records with extra columns are treated as
-      bad records, and if there are too many bad records, an invalid error is
-      returned in the job result. The default value is false. The sourceFormat
-      property determines what BigQuery treats as an extra value: CSV:
-      Trailing columns JSON: Named values that don't match any column names
-    maxBadRecords: [Optional] The maximum number of bad records that BigQuery
-      can ignore when running the job. If the number of bad records exceeds
-      this value, an invalid error is returned in the job result. The default
-      value is 0, which requires that all records are valid.
-    projectionFields: [Experimental] If sourceFormat is set to
-      "DATASTORE_BACKUP", indicates which entity properties to load into
-      BigQuery from a Cloud Datastore backup. Property names are case
-      sensitive and must be top-level properties. If no properties are
-      specified, BigQuery loads all properties. If any named property isn't
-      found in the Cloud Datastore backup, an invalid error is returned in the
-      job result.
-    quote: [Optional] The value that is used to quote data sections in a CSV
-      file. BigQuery converts the string to ISO-8859-1 encoding, and then uses
-      the first byte of the encoded string to split the data in its raw,
-      binary state. The default value is a double-quote ('"'). If your data
-      does not contain quoted sections, set the property value to an empty
-      string. If your data contains quoted newline characters, you must also
-      set the allowQuotedNewlines property to true.
-    schema: [Optional] The schema for the destination table. The schema can be
-      omitted if the destination table already exists, or if you're loading
-      data from Google Cloud Datastore.
-    schemaInline: [Deprecated] The inline schema. For CSV schemas, specify as
-      "Field1:Type1[,Field2:Type2]*". For example, "foo:STRING, bar:INTEGER,
-      baz:FLOAT".
-    schemaInlineFormat: [Deprecated] The format of the schemaInline property.
-    skipLeadingRows: [Optional] The number of rows at the top of a CSV file
-      that BigQuery will skip when loading the data. The default value is 0.
-      This property is useful if you have header rows in the file that should
-      be skipped.
-    sourceFormat: [Optional] The format of the data files. For CSV files,
-      specify "CSV". For datastore backups, specify "DATASTORE_BACKUP". For
-      newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro,
-      specify "AVRO". The default value is CSV.
-    sourceUris: [Required] The fully-qualified URIs that point to your data in
-      Google Cloud Storage. Each URI can contain one '*' wildcard character
-      and it must come after the 'bucket' name.
-    writeDisposition: [Optional] Specifies the action that occurs if the
-      destination table already exists. The following values are supported:
-      WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the
-      table data. WRITE_APPEND: If the table already exists, BigQuery appends
-      the data to the table. WRITE_EMPTY: If the table already exists and
-      contains data, a 'duplicate' error is returned in the job result. The
-      default value is WRITE_APPEND. Each action is atomic and only occurs if
-      BigQuery is able to complete the job successfully. Creation, truncation
-      and append actions occur as one atomic update upon job completion.
-  """
-
-  allowJaggedRows = _messages.BooleanField(1)
-  allowQuotedNewlines = _messages.BooleanField(2)
-  createDisposition = _messages.StringField(3)
-  destinationTable = _messages.MessageField('TableReference', 4)
-  encoding = _messages.StringField(5)
-  fieldDelimiter = _messages.StringField(6)
-  ignoreUnknownValues = _messages.BooleanField(7)
-  maxBadRecords = _messages.IntegerField(8, variant=_messages.Variant.INT32)
-  projectionFields = _messages.StringField(9, repeated=True)
-  quote = _messages.StringField(10, default=u'"')
-  schema = _messages.MessageField('TableSchema', 11)
-  schemaInline = _messages.StringField(12)
-  schemaInlineFormat = _messages.StringField(13)
-  skipLeadingRows = _messages.IntegerField(14, variant=_messages.Variant.INT32)
-  sourceFormat = _messages.StringField(15)
-  sourceUris = _messages.StringField(16, repeated=True)
-  writeDisposition = _messages.StringField(17)
-
-
-class JobConfigurationQuery(_messages.Message):
-  """A JobConfigurationQuery object.
-
-  Messages:
-    TableDefinitionsValue: [Optional] If querying an external data source
-      outside of BigQuery, describes the data format, location and other
-      properties of the data source. By defining these properties, the data
-      source can then be queried as if it were a standard BigQuery table.
-
-  Fields:
-    allowLargeResults: If true, allows the query to produce arbitrarily large
-      result tables at a slight cost in performance. Requires destinationTable
-      to be set.
-    createDisposition: [Optional] Specifies whether the job is allowed to
-      create new tables. The following values are supported: CREATE_IF_NEEDED:
-      If the table does not exist, BigQuery creates the table. CREATE_NEVER:
-      The table must already exist. If it does not, a 'notFound' error is
-      returned in the job result. The default value is CREATE_IF_NEEDED.
-      Creation, truncation and append actions occur as one atomic update upon
-      job completion.
-    defaultDataset: [Optional] Specifies the default dataset to use for
-      unqualified table names in the query.
-    destinationTable: [Optional] Describes the table where the query results
-      should be stored. If not present, a new table will be created to store
-      the results.
-    flattenResults: [Optional] Flattens all nested and repeated fields in the
-      query results. The default value is true. allowLargeResults must be true
-      if this is set to false.
-    maximumBillingTier: [Optional] Limits the billing tier for this job.
-      Queries that have resource usage beyond this tier will fail (without
-      incurring a charge). If unspecified, this will be set to your project
-      default.
-    preserveNulls: [Deprecated] This property is deprecated.
-    priority: [Optional] Specifies a priority for the query. Possible values
-      include INTERACTIVE and BATCH. The default value is INTERACTIVE.
-    query: [Required] BigQuery SQL query to execute.
-    tableDefinitions: [Optional] If querying an external data source outside
-      of BigQuery, describes the data format, location and other properties of
-      the data source. By defining these properties, the data source can then
-      be queried as if it were a standard BigQuery table.
-    useLegacySql: [Experimental] Specifies whether to use BigQuery's legacy
-      SQL dialect for this query. The default value is true. If set to false,
-      the query will use BigQuery's updated SQL dialect with improved
-      standards compliance. When using BigQuery's updated SQL, the values of
-      allowLargeResults and flattenResults are ignored. Queries with
-      useLegacySql set to false will be run as if allowLargeResults is true
-      and flattenResults is false.
-    useQueryCache: [Optional] Whether to look for the result in the query
-      cache. The query cache is a best-effort cache that will be flushed
-      whenever tables in the query are modified. Moreover, the query cache is
-      only available when a query does not have a destination table specified.
-      The default value is true.
-    userDefinedFunctionResources: [Experimental] Describes user-defined
-      function resources used in the query.
-    writeDisposition: [Optional] Specifies the action that occurs if the
-      destination table already exists. The following values are supported:
-      WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the
-      table data. WRITE_APPEND: If the table already exists, BigQuery appends
-      the data to the table. WRITE_EMPTY: If the table already exists and
-      contains data, a 'duplicate' error is returned in the job result. The
-      default value is WRITE_EMPTY. Each action is atomic and only occurs if
-      BigQuery is able to complete the job successfully. Creation, truncation
-      and append actions occur as one atomic update upon job completion.
-  """
-
-  @encoding.MapUnrecognizedFields('additionalProperties')
-  class TableDefinitionsValue(_messages.Message):
-    """[Optional] If querying an external data source outside of BigQuery,
-    describes the data format, location and other properties of the data
-    source. By defining these properties, the data source can then be queried
-    as if it were a standard BigQuery table.
-
-    Messages:
-      AdditionalProperty: An additional property for a TableDefinitionsValue
-        object.
-
-    Fields:
-      additionalProperties: Additional properties of type
-        TableDefinitionsValue
-    """
-
-    class AdditionalProperty(_messages.Message):
-      """An additional property for a TableDefinitionsValue object.
-
-      Fields:
-        key: Name of the additional property.
-        value: A ExternalDataConfiguration attribute.
-      """
-
-      key = _messages.StringField(1)
-      value = _messages.MessageField('ExternalDataConfiguration', 2)
-
-    additionalProperties = _messages.MessageField('AdditionalProperty', 1, repeated=True)
-
-  allowLargeResults = _messages.BooleanField(1)
-  createDisposition = _messages.StringField(2)
-  defaultDataset = _messages.MessageField('DatasetReference', 3)
-  destinationTable = _messages.MessageField('TableReference', 4)
-  flattenResults = _messages.BooleanField(5, default=True)
-  maximumBillingTier = _messages.IntegerField(6, variant=_messages.Variant.INT32, default=1)
-  preserveNulls = _messages.BooleanField(7)
-  priority = _messages.StringField(8)
-  query = _messages.StringField(9)
-  tableDefinitions = _messages.MessageField('TableDefinitionsValue', 10)
-  useLegacySql = _messages.BooleanField(11)
-  useQueryCache = _messages.BooleanField(12, default=True)
-  userDefinedFunctionResources = _messages.MessageField('UserDefinedFunctionResource', 13, repeated=True)
-  writeDisposition = _messages.StringField(14)
-
-
-class JobConfigurationTableCopy(_messages.Message):
-  """A JobConfigurationTableCopy object.
-
-  Fields:
-    createDisposition: [Optional] Specifies whether the job is allowed to
-      create new tables. The following values are supported: CREATE_IF_NEEDED:
-      If the table does not exist, BigQuery creates the table. CREATE_NEVER:
-      The table must already exist. If it does not, a 'notFound' error is
-      returned in the job result. The default value is CREATE_IF_NEEDED.
-      Creation, truncation and append actions occur as one atomic update upon
-      job completion.
-    destinationTable: [Required] The destination table
-    sourceTable: [Pick one] Source table to copy.
-    sourceTables: [Pick one] Source tables to copy.
-    writeDisposition: [Optional] Specifies the action that occurs if the
-      destination table already exists. The following values are supported:
-      WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the
-      table data. WRITE_APPEND: If the table already exists, BigQuery appends
-      the data to the table. WRITE_EMPTY: If the table already exists and
-      contains data, a 'duplicate' error is returned in the job result. The
-      default value is WRITE_EMPTY. Each action is atomic and only occurs if
-      BigQuery is able to complete the job successfully. Creation, truncation
-      and append actions occur as one atomic update upon job completion.
-  """
-
-  createDisposition = _messages.StringField(1)
-  destinationTable = _messages.MessageField('TableReference', 2)
-  sourceTable = _messages.MessageField('TableReference', 3)
-  sourceTables = _messages.MessageField('TableReference', 4, repeated=True)
-  writeDisposition = _messages.StringField(5)
-
-
-class JobList(_messages.Message):
-  """A JobList object.
-
-  Messages:
-    JobsValueListEntry: A JobsValueListEntry object.
-
-  Fields:
-    etag: A hash of this page of results.
-    jobs: List of jobs that were requested.
-    kind: The resource type of the response.
-    nextPageToken: A token to request the next page of results.
-  """
-
-  class JobsValueListEntry(_messages.Message):
-    """A JobsValueListEntry object.
-
-    Fields:
-      configuration: [Full-projection-only] Specifies the job configuration.
-      errorResult: A result object that will be present only if the job has
-        failed.
-      id: Unique opaque ID of the job.
-      jobReference: Job reference uniquely identifying the job.
-      kind: The resource type.
-      state: Running state of the job. When the state is DONE, errorResult can
-        be checked to determine whether the job succeeded or failed.
-      statistics: [Output-only] Information about the job, including starting
-        time and ending time of the job.
-      status: [Full-projection-only] Describes the state of the job.
-      user_email: [Full-projection-only] Email address of the user who ran the
-        job.
-    """
-
-    configuration = _messages.MessageField('JobConfiguration', 1)
-    errorResult = _messages.MessageField('ErrorProto', 2)
-    id = _messages.StringField(3)
-    jobReference = _messages.MessageField('JobReference', 4)
-    kind = _messages.StringField(5, default=u'bigquery#job')
-    state = _messages.StringField(6)
-    statistics = _messages.MessageField('JobStatistics', 7)
-    status = _messages.MessageField('JobStatus', 8)
-    user_email = _messages.StringField(9)
-
-  etag = _messages.StringField(1)
-  jobs = _messages.MessageField('JobsValueListEntry', 2, repeated=True)
-  kind = _messages.StringField(3, default=u'bigquery#jobList')
-  nextPageToken = _messages.StringField(4)
-
-
-class JobReference(_messages.Message):
-  """A JobReference object.
-
-  Fields:
-    jobId: [Required] The ID of the job. The ID must contain only letters
-      (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum
-      length is 1,024 characters.
-    projectId: [Required] The ID of the project containing this job.
-  """
-
-  jobId = _messages.StringField(1)
-  projectId = _messages.StringField(2)
-
-
-class JobStatistics(_messages.Message):
-  """A JobStatistics object.
-
-  Fields:
-    creationTime: [Output-only] Creation time of this job, in milliseconds
-      since the epoch. This field will be present on all jobs.
-    endTime: [Output-only] End time of this job, in milliseconds since the
-      epoch. This field will be present whenever a job is in the DONE state.
-    extract: [Output-only] Statistics for an extract job.
-    load: [Output-only] Statistics for a load job.
-    query: [Output-only] Statistics for a query job.
-    startTime: [Output-only] Start time of this job, in milliseconds since the
-      epoch. This field will be present when the job transitions from the
-      PENDING state to either RUNNING or DONE.
-    totalBytesProcessed: [Output-only] [Deprecated] Use the bytes processed in
-      the query statistics instead.
-  """
-
-  creationTime = _messages.IntegerField(1)
-  endTime = _messages.IntegerField(2)
-  extract = _messages.MessageField('JobStatistics4', 3)
-  load = _messages.MessageField('JobStatistics3', 4)
-  query = _messages.MessageField('JobStatistics2', 5)
-  startTime = _messages.IntegerField(6)
-  totalBytesProcessed = _messages.IntegerField(7)
-
-
-class JobStatistics2(_messages.Message):
-  """A JobStatistics2 object.
-
-  Fields:
-    billingTier: [Output-only] Billing tier for the job.
-    cacheHit: [Output-only] Whether the query result was fetched from the
-      query cache.
-    queryPlan: [Output-only, Experimental] Describes execution plan for the
-      query as a list of stages.
-    referencedTables: [Output-only, Experimental] Referenced tables for the
-      job. Queries that reference more than 50 tables will not have a complete
-      list.
-    totalBytesBilled: [Output-only] Total bytes billed for the job.
-    totalBytesProcessed: [Output-only] Total bytes processed for the job.
-  """
-
-  billingTier = _messages.IntegerField(1, variant=_messages.Variant.INT32)
-  cacheHit = _messages.BooleanField(2)
-  queryPlan = _messages.MessageField('ExplainQueryStage', 3, repeated=True)
-  referencedTables = _messages.MessageField('TableReference', 4, repeated=True)
-  totalBytesBilled = _messages.IntegerField(5)
-  totalBytesProcessed = _messages.IntegerField(6)
-
-
-class JobStatistics3(_messages.Message):
-  """A JobStatistics3 object.
-
-  Fields:
-    inputFileBytes: [Output-only] Number of bytes of source data in a load
-      job.
-    inputFiles: [Output-only] Number of source files in a load job.
-    outputBytes: [Output-only] Size of the loaded data in bytes. Note that
-      while a load job is in the running state, this value may change.
-    outputRows: [Output-only] Number of rows imported in a load job. Note that
-      while an import job is in the running state, this value may change.
-  """
-
-  inputFileBytes = _messages.IntegerField(1)
-  inputFiles = _messages.IntegerField(2)
-  outputBytes = _messages.IntegerField(3)
-  outputRows = _messages.IntegerField(4)
-
-
-class JobStatistics4(_messages.Message):
-  """A JobStatistics4 object.
-
-  Fields:
-    destinationUriFileCounts: [Output-only] Number of files per destination
-      URI or URI pattern specified in the extract configuration. These values
-      will be in the same order as the URIs specified in the 'destinationUris'
-      field.
-  """
-
-  destinationUriFileCounts = _messages.IntegerField(1, repeated=True)
-
-
-class JobStatus(_messages.Message):
-  """A JobStatus object.
-
-  Fields:
-    errorResult: [Output-only] Final error result of the job. If present,
-      indicates that the job has completed and was unsuccessful.
-    errors: [Output-only] All errors encountered during the running of the
-      job. Errors here do not necessarily mean that the job has completed or
-      was unsuccessful.
-    state: [Output-only] Running state of the job.
-  """
-
-  errorResult = _messages.MessageField('ErrorProto', 1)
-  errors = _messages.MessageField('ErrorProto', 2, repeated=True)
-  state = _messages.StringField(3)
-
-
-@encoding.MapUnrecognizedFields('additionalProperties')
-class JsonObject(_messages.Message):
-  """Represents a single JSON object.
-
-  Messages:
-    AdditionalProperty: An additional property for a JsonObject object.
-
-  Fields:
-    additionalProperties: Additional properties of type JsonObject
-  """
-
-  class AdditionalProperty(_messages.Message):
-    """An additional property for a JsonObject object.
-
-    Fields:
-      key: Name of the additional property.
-      value: A JsonValue attribute.
-    """
-
-    key = _messages.StringField(1)
-    value = _messages.MessageField('JsonValue', 2)
-
-  additionalProperties = _messages.MessageField('AdditionalProperty', 1, repeated=True)
-
-
-JsonValue = extra_types.JsonValue
-
-
-class ProjectList(_messages.Message):
-  """A ProjectList object.
-
-  Messages:
-    ProjectsValueListEntry: A ProjectsValueListEntry object.
-
-  Fields:
-    etag: A hash of the page of results
-    kind: The type of list.
-    nextPageToken: A token to request the next page of results.
-    projects: Projects to which you have at least READ access.
-    totalItems: The total number of projects in the list.
-  """
-
-  class ProjectsValueListEntry(_messages.Message):
-    """A ProjectsValueListEntry object.
-
-    Fields:
-      friendlyName: A descriptive name for this project.
-      id: An opaque ID of this project.
-      kind: The resource type.
-      numericId: The numeric ID of this project.
-      projectReference: A unique reference to this project.
-    """
-
-    friendlyName = _messages.StringField(1)
-    id = _messages.StringField(2)
-    kind = _messages.StringField(3, default=u'bigquery#project')
-    numericId = _messages.IntegerField(4, variant=_messages.Variant.UINT64)
-    projectReference = _messages.MessageField('ProjectReference', 5)
-
-  etag = _messages.StringField(1)
-  kind = _messages.StringField(2, default=u'bigquery#projectList')
-  nextPageToken = _messages.StringField(3)
-  projects = _messages.MessageField('ProjectsValueListEntry', 4, repeated=True)
-  totalItems = _messages.IntegerField(5, variant=_messages.Variant.INT32)
-
-
-class ProjectReference(_messages.Message):
-  """A ProjectReference object.
-
-  Fields:
-    projectId: [Required] ID of the project. Can be either the numeric ID or
-      the assigned ID of the project.
-  """
-
-  projectId = _messages.StringField(1)
-
-
-class QueryRequest(_messages.Message):
-  """A QueryRequest object.
-
-  Fields:
-    defaultDataset: [Optional] Specifies the default datasetId and projectId
-      to assume for any unqualified table names in the query. If not set, all
-      table names in the query string must be qualified in the format
-      'datasetId.tableId'.
-    dryRun: [Optional] If set to true, BigQuery doesn't run the job. Instead,
-      if the query is valid, BigQuery returns statistics about the job such as
-      how many bytes would be processed. If the query is invalid, an error
-      returns. The default value is false.
-    kind: The resource type of the request.
-    maxResults: [Optional] The maximum number of rows of data to return per
-      page of results. Setting this flag to a small value such as 1000 and
-      then paging through results might improve reliability when the query
-      result set is large. In addition to this limit, responses are also
-      limited to 10 MB. By default, there is no maximum row count, and only
-      the byte limit applies.
-    preserveNulls: [Deprecated] This property is deprecated.
-    query: [Required] A query string, following the BigQuery query syntax, of
-      the query to execute. Example: "SELECT count(f1) FROM
-      [myProjectId:myDatasetId.myTableId]".
-    timeoutMs: [Optional] How long to wait for the query to complete, in
-      milliseconds, before the request times out and returns. Note that this
-      is only a timeout for the request, not the query. If the query takes
-      longer to run than the timeout value, the call returns without any
-      results and with the 'jobComplete' flag set to false. You can call
-      GetQueryResults() to wait for the query to complete and read the
-      results. The default value is 10000 milliseconds (10 seconds).
-    useLegacySql: [Experimental] Specifies whether to use BigQuery's legacy
-      SQL dialect for this query. The default value is true. If set to false,
-      the query will use BigQuery's updated SQL dialect with improved
-      standards compliance. When using BigQuery's updated SQL, the values of
-      allowLargeResults and flattenResults are ignored. Queries with
-      useLegacySql set to false will be run as if allowLargeResults is true
-      and flattenResults is false.
-    useQueryCache: [Optional] Whether to look for the result in the query
-      cache. The query cache is a best-effort cache that will be flushed
-      whenever tables in the query are modified. The default value is true.
-  """
-
-  defaultDataset = _messages.MessageField('DatasetReference', 1)
-  dryRun = _messages.BooleanField(2)
-  kind = _messages.StringField(3, default=u'bigquery#queryRequest')
-  maxResults = _messages.IntegerField(4, variant=_messages.Variant.UINT32)
-  preserveNulls = _messages.BooleanField(5)
-  query = _messages.StringField(6)
-  timeoutMs = _messages.IntegerField(7, variant=_messages.Variant.UINT32)
-  useLegacySql = _messages.BooleanField(8)
-  useQueryCache = _messages.BooleanField(9, default=True)
-
-
-class QueryResponse(_messages.Message):
-  """A QueryResponse object.
-
-  Fields:
-    cacheHit: Whether the query result was fetched from the query cache.
-    errors: [Output-only] All errors and warnings encountered during the
-      running of the job. Errors here do not necessarily mean that the job has
-      completed or was unsuccessful.
-    jobComplete: Whether the query has completed or not. If rows or totalRows
-      are present, this will always be true. If this is false, totalRows will
-      not be available.
-    jobReference: Reference to the Job that was created to run the query. This
-      field will be present even if the original request timed out, in which
-      case GetQueryResults can be used to read the results once the query has
-      completed. Since this API only returns the first page of results,
-      subsequent pages can be fetched via the same mechanism
-      (GetQueryResults).
-    kind: The resource type.
-    pageToken: A token used for paging results.
-    rows: An object with as many results as can be contained within the
-      maximum permitted reply size. To get any additional rows, you can call
-      GetQueryResults and specify the jobReference returned above.
-    schema: The schema of the results. Present only when the query completes
-      successfully.
-    totalBytesProcessed: The total number of bytes processed for this query.
-      If this query was a dry run, this is the number of bytes that would be
-      processed if the query were run.
-    totalRows: The total number of rows in the complete query result set,
-      which can be more than the number of rows in this single page of
-      results.
-  """
-
-  cacheHit = _messages.BooleanField(1)
-  errors = _messages.MessageField('ErrorProto', 2, repeated=True)
-  jobComplete = _messages.BooleanField(3)
-  jobReference = _messages.MessageField('JobReference', 4)
-  kind = _messages.StringField(5, default=u'bigquery#queryResponse')
-  pageToken = _messages.StringField(6)
-  rows = _messages.MessageField('TableRow', 7, repeated=True)
-  schema = _messages.MessageField('TableSchema', 8)
-  totalBytesProcessed = _messages.IntegerField(9)
-  totalRows = _messages.IntegerField(10, variant=_messages.Variant.UINT64)
-
-
-class StandardQueryParameters(_messages.Message):
-  """Query parameters accepted by all methods.
-
-  Enums:
-    AltValueValuesEnum: Data format for the response.
-
-  Fields:
-    alt: Data format for the response.
-    fields: Selector specifying which fields to include in a partial response.
-    key: API key. Your API key identifies your project and provides you with
-      API access, quota, and reports. Required unless you provide an OAuth 2.0
-      token.
-    oauth_token: OAuth 2.0 token for the current user.
-    prettyPrint: Returns response with indentations and line breaks.
-    quotaUser: Available to use for quota purposes for server-side
-      applications. Can be any arbitrary string assigned to a user, but should
-      not exceed 40 characters. Overrides userIp if both are provided.
-    trace: A tracing token of the form "token:<tokenid>" to include in api
-      requests.
-    userIp: IP address of the site where the request originates. Use this if
-      you want to enforce per-user limits.
-  """
-
-  class AltValueValuesEnum(_messages.Enum):
-    """Data format for the response.
-
-    Values:
-      csv: Responses with Content-Type of text/csv
-      json: Responses with Content-Type of application/json
-    """
-    csv = 0
-    json = 1
-
-  alt = _messages.EnumField('AltValueValuesEnum', 1, default=u'json')
-  fields = _messages.StringField(2)
-  key = _messages.StringField(3)
-  oauth_token = _messages.StringField(4)
-  prettyPrint = _messages.BooleanField(5, default=True)
-  quotaUser = _messages.StringField(6)
-  trace = _messages.StringField(7)
-  userIp = _messages.StringField(8)
-
-
-class Streamingbuffer(_messages.Message):
-  """A Streamingbuffer object.
-
-  Fields:
-    estimatedBytes: [Output-only] A lower-bound estimate of the number of
-      bytes currently in the streaming buffer.
-    estimatedRows: [Output-only] A lower-bound estimate of the number of rows
-      currently in the streaming buffer.
-    oldestEntryTime: [Output-only] Contains the timestamp of the oldest entry
-      in the streaming buffer, in milliseconds since the epoch, if the
-      streaming buffer is available.
-  """
-
-  estimatedBytes = _messages.IntegerField(1, variant=_messages.Variant.UINT64)
-  estimatedRows = _messages.IntegerField(2, variant=_messages.Variant.UINT64)
-  oldestEntryTime = _messages.IntegerField(3, variant=_messages.Variant.UINT64)
-
-
-class Table(_messages.Message):
-  """A Table object.
-
-  Fields:
-    creationTime: [Output-only] The time when this table was created, in
-      milliseconds since the epoch.
-    description: [Optional] A user-friendly description of this table.
-    etag: [Output-only] A hash of this resource.
-    expirationTime: [Optional] The time when this table expires, in
-      milliseconds since the epoch. If not present, the table will persist
-      indefinitely. Expired tables will be deleted and their storage
-      reclaimed.
-    externalDataConfiguration: [Optional] Describes the data format, location,
-      and other properties of a table stored outside of BigQuery. By defining
-      these properties, the data source can then be queried as if it were a
-      standard BigQuery table.
-    friendlyName: [Optional] A descriptive name for this table.
-    id: [Output-only] An opaque ID uniquely identifying the table.
-    kind: [Output-only] The type of the resource.
-    lastModifiedTime: [Output-only] The time when this table was last
-      modified, in milliseconds since the epoch.
-    location: [Output-only] The geographic location where the table resides.
-      This value is inherited from the dataset.
-    numBytes: [Output-only] The size of this table in bytes, excluding any
-      data in the streaming buffer.
-    numRows: [Output-only] The number of rows of data in this table, excluding
-      any data in the streaming buffer.
-    partitionConfigurations: [Experimental] List of partition configurations
-      for this table. Currently only one configuration can be specified and it
-      can only be an interval partition with type daily.
-    schema: [Optional] Describes the schema of this table.
-    selfLink: [Output-only] A URL that can be used to access this resource
-      again.
-    streamingBuffer: [Output-only] Contains information regarding this table's
-      streaming buffer, if one is present. This field will be absent if the
-      table is not being streamed to or if there is no data in the streaming
-      buffer.
-    tableReference: [Required] Reference describing the ID of this table.
-    type: [Output-only] Describes the table type. The following values are
-      supported: TABLE: A normal BigQuery table. VIEW: A virtual table defined
-      by a SQL query. EXTERNAL: A table that references data stored in an
-      external storage system, such as Google Cloud Storage. The default value
-      is TABLE.
-    view: [Optional] The view definition.
-  """
-
-  creationTime = _messages.IntegerField(1)
-  description = _messages.StringField(2)
-  etag = _messages.StringField(3)
-  expirationTime = _messages.IntegerField(4)
-  externalDataConfiguration = _messages.MessageField('ExternalDataConfiguration', 5)
-  friendlyName = _messages.StringField(6)
-  id = _messages.StringField(7)
-  kind = _messages.StringField(8, default=u'bigquery#table')
-  lastModifiedTime = _messages.IntegerField(9, variant=_messages.Variant.UINT64)
-  location = _messages.StringField(10)
-  numBytes = _messages.IntegerField(11)
-  numRows = _messages.IntegerField(12, variant=_messages.Variant.UINT64)
-  partitionConfigurations = _messages.MessageField('TablePartitionConfiguration', 13, repeated=True)
-  schema = _messages.MessageField('TableSchema', 14)
-  selfLink = _messages.StringField(15)
-  streamingBuffer = _messages.MessageField('Streamingbuffer', 16)
-  tableReference = _messages.MessageField('TableReference', 17)
-  type = _messages.StringField(18)
-  view = _messages.MessageField('ViewDefinition', 19)
-
-
-class TableCell(_messages.Message):
-  """A TableCell object.
-
-  Fields:
-    v: A extra_types.JsonValue attribute.
-  """
-
-  v = _messages.MessageField('extra_types.JsonValue', 1)
-
-
-class TableDataInsertAllRequest(_messages.Message):
-  """A TableDataInsertAllRequest object.
-
-  Messages:
-    RowsValueListEntry: A RowsValueListEntry object.
-
-  Fields:
-    ignoreUnknownValues: [Optional] Accept rows that contain values that do
-      not match the schema. The unknown values are ignored. Default is false,
-      which treats unknown values as errors.
-    kind: The resource type of the response.
-    rows: The rows to insert.
-    skipInvalidRows: [Optional] Insert all valid rows of a request, even if
-      invalid rows exist. The default value is false, which causes the entire
-      request to fail if any invalid rows exist.
-    templateSuffix: [Experimental] If specified, treats the destination table
-      as a base template, and inserts the rows into an instance table named
-      "{destination}{templateSuffix}". BigQuery will manage creation of the
-      instance table, using the schema of the base template table. See
-      https://cloud.google.com/bigquery/streaming-data-into-bigquery#template-
-      tables for considerations when working with templates tables.
-  """
-
-  class RowsValueListEntry(_messages.Message):
-    """A RowsValueListEntry object.
-
-    Fields:
-      insertId: [Optional] A unique ID for each row. BigQuery uses this
-        property to detect duplicate insertion requests on a best-effort
-        basis.
-      json: [Required] A JSON object that contains a row of data. The object's
-        properties and values must match the destination table's schema.
-    """
-
-    insertId = _messages.StringField(1)
-    json = _messages.MessageField('JsonObject', 2)
-
-  ignoreUnknownValues = _messages.BooleanField(1)
-  kind = _messages.StringField(2, default=u'bigquery#tableDataInsertAllRequest')
-  rows = _messages.MessageField('RowsValueListEntry', 3, repeated=True)
-  skipInvalidRows = _messages.BooleanField(4)
-  templateSuffix = _messages.StringField(5)
-
-
-class TableDataInsertAllResponse(_messages.Message):
-  """A TableDataInsertAllResponse object.
-
-  Messages:
-    InsertErrorsValueListEntry: A InsertErrorsValueListEntry object.
-
-  Fields:
-    insertErrors: An array of errors for rows that were not inserted.
-    kind: The resource type of the response.
-  """
-
-  class InsertErrorsValueListEntry(_messages.Message):
-    """A InsertErrorsValueListEntry object.
-
-    Fields:
-      errors: Error information for the row indicated by the index property.
-      index: The index of the row that error applies to.
-    """
-
-    errors = _messages.MessageField('ErrorProto', 1, repeated=True)
-    index = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
-
-  insertErrors = _messages.MessageField('InsertErrorsValueListEntry', 1, repeated=True)
-  kind = _messages.StringField(2, default=u'bigquery#tableDataInsertAllResponse')
-
-
-class TableDataList(_messages.Message):
-  """A TableDataList object.
-
-  Fields:
-    etag: A hash of this page of results.
-    kind: The resource type of the response.
-    pageToken: A token used for paging results. Providing this token instead
-      of the startIndex parameter can help you retrieve stable results when an
-      underlying table is changing.
-    rows: Rows of results.
-    totalRows: The total number of rows in the complete table.
-  """
-
-  etag = _messages.StringField(1)
-  kind = _messages.StringField(2, default=u'bigquery#tableDataList')
-  pageToken = _messages.StringField(3)
-  rows = _messages.MessageField('TableRow', 4, repeated=True)
-  totalRows = _messages.IntegerField(5)
-
-
-class TableFieldSchema(_messages.Message):
-  """A TableFieldSchema object.
-
-  Fields:
-    description: [Optional] The field description. The maximum length is 16K
-      characters.
-    fields: [Optional] Describes the nested schema fields if the type property
-      is set to RECORD.
-    mode: [Optional] The field mode. Possible values include NULLABLE,
-      REQUIRED and REPEATED. The default value is NULLABLE.
-    name: [Required] The field name. The name must contain only letters (a-z,
-      A-Z), numbers (0-9), or underscores (_), and must start with a letter or
-      underscore. The maximum length is 128 characters.
-    type: [Required] The field data type. Possible values include STRING,
-      BYTES, INTEGER, FLOAT, BOOLEAN, TIMESTAMP or RECORD (where RECORD
-      indicates that the field contains a nested schema).
-  """
-
-  description = _messages.StringField(1)
-  fields = _messages.MessageField('TableFieldSchema', 2, repeated=True)
-  mode = _messages.StringField(3)
-  name = _messages.StringField(4)
-  type = _messages.StringField(5)
-
-
-class TableList(_messages.Message):
-  """A TableList object.
-
-  Messages:
-    TablesValueListEntry: A TablesValueListEntry object.
-
-  Fields:
-    etag: A hash of this page of results.
-    kind: The type of list.
-    nextPageToken: A token to request the next page of results.
-    tables: Tables in the requested dataset.
-    totalItems: The total number of tables in the dataset.
-  """
-
-  class TablesValueListEntry(_messages.Message):
-    """A TablesValueListEntry object.
-
-    Fields:
-      friendlyName: The user-friendly name for this table.
-      id: An opaque ID of the table
-      kind: The resource type.
-      tableReference: A reference uniquely identifying the table.
-      type: The type of table. Possible values are: TABLE, VIEW.
-    """
-
-    friendlyName = _messages.StringField(1)
-    id = _messages.StringField(2)
-    kind = _messages.StringField(3, default=u'bigquery#table')
-    tableReference = _messages.MessageField('TableReference', 4)
-    type = _messages.StringField(5)
-
-  etag = _messages.StringField(1)
-  kind = _messages.StringField(2, default=u'bigquery#tableList')
-  nextPageToken = _messages.StringField(3)
-  tables = _messages.MessageField('TablesValueListEntry', 4, repeated=True)
-  totalItems = _messages.IntegerField(5, variant=_messages.Variant.INT32)
-
-
-class TablePartitionConfiguration(_messages.Message):
-  """[Required] A partition configuration. Only one type of partition should
-  be configured.
-
-  Fields:
-    interval: [Pick one] Configures an interval partition.
-  """
-
-  interval = _messages.MessageField('IntervalPartitionConfiguration', 1)
-
-
-class TableReference(_messages.Message):
-  """A TableReference object.
-
-  Fields:
-    datasetId: [Required] The ID of the dataset containing this table.
-    projectId: [Required] The ID of the project containing this table.
-    tableId: [Required] The ID of the table. The ID must contain only letters
-      (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is
-      1,024 characters.
-  """
-
-  datasetId = _messages.StringField(1)
-  projectId = _messages.StringField(2)
-  tableId = _messages.StringField(3)
-
-
-class TableRow(_messages.Message):
-  """A TableRow object.
-
-  Fields:
-    f: Represents a single row in the result set, consisting of one or more
-      fields.
-  """
-
-  f = _messages.MessageField('TableCell', 1, repeated=True)
-
-
-class TableSchema(_messages.Message):
-  """A TableSchema object.
-
-  Fields:
-    fields: Describes the fields in a table.
-  """
-
-  fields = _messages.MessageField('TableFieldSchema', 1, repeated=True)
-
-
-class UserDefinedFunctionResource(_messages.Message):
-  """A UserDefinedFunctionResource object.
-
-  Fields:
-    inlineCode: [Pick one] An inline resource that contains code for a user-
-      defined function (UDF). Providing a inline code resource is equivalent
-      to providing a URI for a file containing the same code.
-    resourceUri: [Pick one] A code resource to load from a Google Cloud
-      Storage URI (gs://bucket/path).
-  """
-
-  inlineCode = _messages.StringField(1)
-  resourceUri = _messages.StringField(2)
-
-
-class ViewDefinition(_messages.Message):
-  """A ViewDefinition object.
-
-  Fields:
-    query: [Required] A query that BigQuery executes when the view is
-      referenced.
-    userDefinedFunctionResources: [Experimental] Describes user-defined
-      function resources used in the query.
-  """
-
-  query = _messages.StringField(1)
-  userDefinedFunctionResources = _messages.MessageField('UserDefinedFunctionResource', 2, repeated=True)
-
-

http://git-wip-us.apache.org/repos/asf/beam/blob/21925496/sdks/python/apache_beam/internal/clients/storage/__init__.py
----------------------------------------------------------------------
diff --git a/sdks/python/apache_beam/internal/clients/storage/__init__.py b/sdks/python/apache_beam/internal/clients/storage/__init__.py
deleted file mode 100644
index e4650e8..0000000
--- a/sdks/python/apache_beam/internal/clients/storage/__init__.py
+++ /dev/null
@@ -1,27 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one or more
-# contributor license agreements.  See the NOTICE file distributed with
-# this work for additional information regarding copyright ownership.
-# The ASF licenses this file to You under the Apache License, Version 2.0
-# (the "License"); you may not use this file except in compliance with
-# the License.  You may obtain a copy of the License at
-#
-#    http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-"""Common imports for generated storage client library."""
-# pylint:disable=wildcard-import
-
-import pkgutil
-
-from apitools.base.py import *
-from apache_beam.internal.clients.storage.storage_v1_client import *
-from apache_beam.internal.clients.storage.storage_v1_messages import *
-
-__path__ = pkgutil.extend_path(__path__, __name__)


Mime
View raw message