beam-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From al...@apache.org
Subject [17/22] beam git commit: Rename google_cloud_dataflow and google_cloud_platform
Date Thu, 23 Feb 2017 01:23:12 GMT
http://git-wip-us.apache.org/repos/asf/beam/blob/59ad58ac/sdks/python/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_messages.py
----------------------------------------------------------------------
diff --git a/sdks/python/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_messages.py b/sdks/python/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_messages.py
new file mode 100644
index 0000000..4045428
--- /dev/null
+++ b/sdks/python/apache_beam/io/gcp/internal/clients/bigquery/bigquery_v2_messages.py
@@ -0,0 +1,1910 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#    http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+#
+
+"""Generated message classes for bigquery version v2.
+
+A data platform for customers to create, manage, share and query data.
+"""
+# NOTE: This file is autogenerated and should not be edited by hand.
+
+from apitools.base.protorpclite import messages as _messages
+from apitools.base.py import encoding
+from apitools.base.py import extra_types
+
+
+package = 'bigquery'
+
+
+class BigqueryDatasetsDeleteRequest(_messages.Message):
+  """A BigqueryDatasetsDeleteRequest object.
+
+  Fields:
+    datasetId: Dataset ID of dataset being deleted
+    deleteContents: If True, delete all the tables in the dataset. If False
+      and the dataset contains tables, the request will fail. Default is False
+    projectId: Project ID of the dataset being deleted
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  deleteContents = _messages.BooleanField(2)
+  projectId = _messages.StringField(3, required=True)
+
+
+class BigqueryDatasetsDeleteResponse(_messages.Message):
+  """An empty BigqueryDatasetsDelete response."""
+
+
+class BigqueryDatasetsGetRequest(_messages.Message):
+  """A BigqueryDatasetsGetRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the requested dataset
+    projectId: Project ID of the requested dataset
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+
+
+class BigqueryDatasetsInsertRequest(_messages.Message):
+  """A BigqueryDatasetsInsertRequest object.
+
+  Fields:
+    dataset: A Dataset resource to be passed as the request body.
+    projectId: Project ID of the new dataset
+  """
+
+  dataset = _messages.MessageField('Dataset', 1)
+  projectId = _messages.StringField(2, required=True)
+
+
+class BigqueryDatasetsListRequest(_messages.Message):
+  """A BigqueryDatasetsListRequest object.
+
+  Fields:
+    all: Whether to list all datasets, including hidden ones
+    maxResults: The maximum number of results to return
+    pageToken: Page token, returned by a previous call, to request the next
+      page of results
+    projectId: Project ID of the datasets to be listed
+  """
+
+  all = _messages.BooleanField(1)
+  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
+  pageToken = _messages.StringField(3)
+  projectId = _messages.StringField(4, required=True)
+
+
+class BigqueryDatasetsPatchRequest(_messages.Message):
+  """A BigqueryDatasetsPatchRequest object.
+
+  Fields:
+    dataset: A Dataset resource to be passed as the request body.
+    datasetId: Dataset ID of the dataset being updated
+    projectId: Project ID of the dataset being updated
+  """
+
+  dataset = _messages.MessageField('Dataset', 1)
+  datasetId = _messages.StringField(2, required=True)
+  projectId = _messages.StringField(3, required=True)
+
+
+class BigqueryDatasetsUpdateRequest(_messages.Message):
+  """A BigqueryDatasetsUpdateRequest object.
+
+  Fields:
+    dataset: A Dataset resource to be passed as the request body.
+    datasetId: Dataset ID of the dataset being updated
+    projectId: Project ID of the dataset being updated
+  """
+
+  dataset = _messages.MessageField('Dataset', 1)
+  datasetId = _messages.StringField(2, required=True)
+  projectId = _messages.StringField(3, required=True)
+
+
+class BigqueryJobsCancelRequest(_messages.Message):
+  """A BigqueryJobsCancelRequest object.
+
+  Fields:
+    jobId: [Required] Job ID of the job to cancel
+    projectId: [Required] Project ID of the job to cancel
+  """
+
+  jobId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+
+
+class BigqueryJobsGetQueryResultsRequest(_messages.Message):
+  """A BigqueryJobsGetQueryResultsRequest object.
+
+  Fields:
+    jobId: [Required] Job ID of the query job
+    maxResults: Maximum number of results to read
+    pageToken: Page token, returned by a previous call, to request the next
+      page of results
+    projectId: [Required] Project ID of the query job
+    startIndex: Zero-based index of the starting row
+    timeoutMs: How long to wait for the query to complete, in milliseconds,
+      before returning. Default is 10 seconds. If the timeout passes before
+      the job completes, the 'jobComplete' field in the response will be false
+  """
+
+  jobId = _messages.StringField(1, required=True)
+  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
+  pageToken = _messages.StringField(3)
+  projectId = _messages.StringField(4, required=True)
+  startIndex = _messages.IntegerField(5, variant=_messages.Variant.UINT64)
+  timeoutMs = _messages.IntegerField(6, variant=_messages.Variant.UINT32)
+
+
+class BigqueryJobsGetRequest(_messages.Message):
+  """A BigqueryJobsGetRequest object.
+
+  Fields:
+    jobId: [Required] Job ID of the requested job
+    projectId: [Required] Project ID of the requested job
+  """
+
+  jobId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+
+
+class BigqueryJobsInsertRequest(_messages.Message):
+  """A BigqueryJobsInsertRequest object.
+
+  Fields:
+    job: A Job resource to be passed as the request body.
+    projectId: Project ID of the project that will be billed for the job
+  """
+
+  job = _messages.MessageField('Job', 1)
+  projectId = _messages.StringField(2, required=True)
+
+
+class BigqueryJobsListRequest(_messages.Message):
+  """A BigqueryJobsListRequest object.
+
+  Enums:
+    ProjectionValueValuesEnum: Restrict information returned to a set of
+      selected fields
+    StateFilterValueValuesEnum: Filter for job state
+
+  Fields:
+    allUsers: Whether to display jobs owned by all users in the project.
+      Default false
+    maxResults: Maximum number of results to return
+    pageToken: Page token, returned by a previous call, to request the next
+      page of results
+    projectId: Project ID of the jobs to list
+    projection: Restrict information returned to a set of selected fields
+    stateFilter: Filter for job state
+  """
+
+  class ProjectionValueValuesEnum(_messages.Enum):
+    """Restrict information returned to a set of selected fields
+
+    Values:
+      full: Includes all job data
+      minimal: Does not include the job configuration
+    """
+    full = 0
+    minimal = 1
+
+  class StateFilterValueValuesEnum(_messages.Enum):
+    """Filter for job state
+
+    Values:
+      done: Finished jobs
+      pending: Pending jobs
+      running: Running jobs
+    """
+    done = 0
+    pending = 1
+    running = 2
+
+  allUsers = _messages.BooleanField(1)
+  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
+  pageToken = _messages.StringField(3)
+  projectId = _messages.StringField(4, required=True)
+  projection = _messages.EnumField('ProjectionValueValuesEnum', 5)
+  stateFilter = _messages.EnumField('StateFilterValueValuesEnum', 6, repeated=True)
+
+
+class BigqueryJobsQueryRequest(_messages.Message):
+  """A BigqueryJobsQueryRequest object.
+
+  Fields:
+    projectId: Project ID of the project billed for the query
+    queryRequest: A QueryRequest resource to be passed as the request body.
+  """
+
+  projectId = _messages.StringField(1, required=True)
+  queryRequest = _messages.MessageField('QueryRequest', 2)
+
+
+class BigqueryProjectsListRequest(_messages.Message):
+  """A BigqueryProjectsListRequest object.
+
+  Fields:
+    maxResults: Maximum number of results to return
+    pageToken: Page token, returned by a previous call, to request the next
+      page of results
+  """
+
+  maxResults = _messages.IntegerField(1, variant=_messages.Variant.UINT32)
+  pageToken = _messages.StringField(2)
+
+
+class BigqueryTabledataInsertAllRequest(_messages.Message):
+  """A BigqueryTabledataInsertAllRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the destination table.
+    projectId: Project ID of the destination table.
+    tableDataInsertAllRequest: A TableDataInsertAllRequest resource to be
+      passed as the request body.
+    tableId: Table ID of the destination table.
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+  tableDataInsertAllRequest = _messages.MessageField('TableDataInsertAllRequest', 3)
+  tableId = _messages.StringField(4, required=True)
+
+
+class BigqueryTabledataListRequest(_messages.Message):
+  """A BigqueryTabledataListRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the table to read
+    maxResults: Maximum number of results to return
+    pageToken: Page token, returned by a previous call, identifying the result
+      set
+    projectId: Project ID of the table to read
+    startIndex: Zero-based index of the starting row to read
+    tableId: Table ID of the table to read
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
+  pageToken = _messages.StringField(3)
+  projectId = _messages.StringField(4, required=True)
+  startIndex = _messages.IntegerField(5, variant=_messages.Variant.UINT64)
+  tableId = _messages.StringField(6, required=True)
+
+
+class BigqueryTablesDeleteRequest(_messages.Message):
+  """A BigqueryTablesDeleteRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the table to delete
+    projectId: Project ID of the table to delete
+    tableId: Table ID of the table to delete
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+  tableId = _messages.StringField(3, required=True)
+
+
+class BigqueryTablesDeleteResponse(_messages.Message):
+  """An empty BigqueryTablesDelete response."""
+
+
+class BigqueryTablesGetRequest(_messages.Message):
+  """A BigqueryTablesGetRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the requested table
+    projectId: Project ID of the requested table
+    tableId: Table ID of the requested table
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+  tableId = _messages.StringField(3, required=True)
+
+
+class BigqueryTablesInsertRequest(_messages.Message):
+  """A BigqueryTablesInsertRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the new table
+    projectId: Project ID of the new table
+    table: A Table resource to be passed as the request body.
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+  table = _messages.MessageField('Table', 3)
+
+
+class BigqueryTablesListRequest(_messages.Message):
+  """A BigqueryTablesListRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the tables to list
+    maxResults: Maximum number of results to return
+    pageToken: Page token, returned by a previous call, to request the next
+      page of results
+    projectId: Project ID of the tables to list
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  maxResults = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
+  pageToken = _messages.StringField(3)
+  projectId = _messages.StringField(4, required=True)
+
+
+class BigqueryTablesPatchRequest(_messages.Message):
+  """A BigqueryTablesPatchRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the table to update
+    projectId: Project ID of the table to update
+    table: A Table resource to be passed as the request body.
+    tableId: Table ID of the table to update
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+  table = _messages.MessageField('Table', 3)
+  tableId = _messages.StringField(4, required=True)
+
+
+class BigqueryTablesUpdateRequest(_messages.Message):
+  """A BigqueryTablesUpdateRequest object.
+
+  Fields:
+    datasetId: Dataset ID of the table to update
+    projectId: Project ID of the table to update
+    table: A Table resource to be passed as the request body.
+    tableId: Table ID of the table to update
+  """
+
+  datasetId = _messages.StringField(1, required=True)
+  projectId = _messages.StringField(2, required=True)
+  table = _messages.MessageField('Table', 3)
+  tableId = _messages.StringField(4, required=True)
+
+
+class BigtableColumn(_messages.Message):
+  """A BigtableColumn object.
+
+  Fields:
+    encoding: [Optional] The encoding of the values when the type is not
+      STRING. Acceptable encoding values are: TEXT - indicates values are
+      alphanumeric text strings. BINARY - indicates values are encoded using
+      HBase Bytes.toBytes family of functions. 'encoding' can also be set at
+      the column family level. However, the setting at this level takes
+      precedence if 'encoding' is set at both levels.
+    fieldName: [Optional] If the qualifier is not a valid BigQuery field
+      identifier i.e. does not match [a-zA-Z][a-zA-Z0-9_]*, a valid identifier
+      must be provided as the column field name and is used as field name in
+      queries.
+    onlyReadLatest: [Optional] If this is set, only the latest version of
+      value in this column are exposed. 'onlyReadLatest' can also be set at
+      the column family level. However, the setting at this level takes
+      precedence if 'onlyReadLatest' is set at both levels.
+    qualifierEncoded: [Required] Qualifier of the column. Columns in the
+      parent column family that has this exact qualifier are exposed as .
+      field. If the qualifier is valid UTF-8 string, it can be specified in
+      the qualifier_string field. Otherwise, a base-64 encoded value must be
+      set to qualifier_encoded. The column field name is the same as the
+      column qualifier. However, if the qualifier is not a valid BigQuery
+      field identifier i.e. does not match [a-zA-Z][a-zA-Z0-9_]*, a valid
+      identifier must be provided as field_name.
+    qualifierString: A string attribute.
+    type: [Optional] The type to convert the value in cells of this column.
+      The values are expected to be encoded using HBase Bytes.toBytes function
+      when using the BINARY encoding value. Following BigQuery types are
+      allowed (case-sensitive) - BYTES STRING INTEGER FLOAT BOOLEAN Defaut
+      type is BYTES. 'type' can also be set at the column family level.
+      However, the setting at this level takes precedence if 'type' is set at
+      both levels.
+  """
+
+  encoding = _messages.StringField(1)
+  fieldName = _messages.StringField(2)
+  onlyReadLatest = _messages.BooleanField(3)
+  qualifierEncoded = _messages.BytesField(4)
+  qualifierString = _messages.StringField(5)
+  type = _messages.StringField(6)
+
+
+class BigtableColumnFamily(_messages.Message):
+  """A BigtableColumnFamily object.
+
+  Fields:
+    columns: [Optional] Lists of columns that should be exposed as individual
+      fields as opposed to a list of (column name, value) pairs. All columns
+      whose qualifier matches a qualifier in this list can be accessed as ..
+      Other columns can be accessed as a list through .Column field.
+    encoding: [Optional] The encoding of the values when the type is not
+      STRING. Acceptable encoding values are: TEXT - indicates values are
+      alphanumeric text strings. BINARY - indicates values are encoded using
+      HBase Bytes.toBytes family of functions. This can be overridden for a
+      specific column by listing that column in 'columns' and specifying an
+      encoding for it.
+    familyId: Identifier of the column family.
+    onlyReadLatest: [Optional] If this is set only the latest version of value
+      are exposed for all columns in this column family. This can be
+      overridden for a specific column by listing that column in 'columns' and
+      specifying a different setting for that column.
+    type: [Optional] The type to convert the value in cells of this column
+      family. The values are expected to be encoded using HBase Bytes.toBytes
+      function when using the BINARY encoding value. Following BigQuery types
+      are allowed (case-sensitive) - BYTES STRING INTEGER FLOAT BOOLEAN Defaut
+      type is BYTES. This can be overridden for a specific column by listing
+      that column in 'columns' and specifying a type for it.
+  """
+
+  columns = _messages.MessageField('BigtableColumn', 1, repeated=True)
+  encoding = _messages.StringField(2)
+  familyId = _messages.StringField(3)
+  onlyReadLatest = _messages.BooleanField(4)
+  type = _messages.StringField(5)
+
+
+class BigtableOptions(_messages.Message):
+  """A BigtableOptions object.
+
+  Fields:
+    columnFamilies: [Optional] List of column families to expose in the table
+      schema along with their types. This list restricts the column families
+      that can be referenced in queries and specifies their value types. You
+      can use this list to do type conversions - see the 'type' field for more
+      details. If you leave this list empty, all column families are present
+      in the table schema and their values are read as BYTES. During a query
+      only the column families referenced in that query are read from
+      Bigtable.
+    ignoreUnspecifiedColumnFamilies: [Optional] If field is true, then the
+      column families that are not specified in columnFamilies list are not
+      exposed in the table schema. Otherwise, they are read with BYTES type
+      values. The default value is false.
+  """
+
+  columnFamilies = _messages.MessageField('BigtableColumnFamily', 1, repeated=True)
+  ignoreUnspecifiedColumnFamilies = _messages.BooleanField(2)
+
+
+class CsvOptions(_messages.Message):
+  """A CsvOptions object.
+
+  Fields:
+    allowJaggedRows: [Optional] Indicates if BigQuery should accept rows that
+      are missing trailing optional columns. If true, BigQuery treats missing
+      trailing columns as null values. If false, records with missing trailing
+      columns are treated as bad records, and if there are too many bad
+      records, an invalid error is returned in the job result. The default
+      value is false.
+    allowQuotedNewlines: [Optional] Indicates if BigQuery should allow quoted
+      data sections that contain newline characters in a CSV file. The default
+      value is false.
+    encoding: [Optional] The character encoding of the data. The supported
+      values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery
+      decodes the data after the raw, binary data has been split using the
+      values of the quote and fieldDelimiter properties.
+    fieldDelimiter: [Optional] The separator for fields in a CSV file.
+      BigQuery converts the string to ISO-8859-1 encoding, and then uses the
+      first byte of the encoded string to split the data in its raw, binary
+      state. BigQuery also supports the escape sequence "\t" to specify a tab
+      separator. The default value is a comma (',').
+    quote: [Optional] The value that is used to quote data sections in a CSV
+      file. BigQuery converts the string to ISO-8859-1 encoding, and then uses
+      the first byte of the encoded string to split the data in its raw,
+      binary state. The default value is a double-quote ('"'). If your data
+      does not contain quoted sections, set the property value to an empty
+      string. If your data contains quoted newline characters, you must also
+      set the allowQuotedNewlines property to true.
+    skipLeadingRows: [Optional] The number of rows at the top of a CSV file
+      that BigQuery will skip when reading the data. The default value is 0.
+      This property is useful if you have header rows in the file that should
+      be skipped.
+  """
+
+  allowJaggedRows = _messages.BooleanField(1)
+  allowQuotedNewlines = _messages.BooleanField(2)
+  encoding = _messages.StringField(3)
+  fieldDelimiter = _messages.StringField(4)
+  quote = _messages.StringField(5, default=u'"')
+  skipLeadingRows = _messages.IntegerField(6, variant=_messages.Variant.INT32)
+
+
+class Dataset(_messages.Message):
+  """A Dataset object.
+
+  Messages:
+    AccessValueListEntry: A AccessValueListEntry object.
+
+  Fields:
+    access: [Optional] An array of objects that define dataset access for one
+      or more entities. You can set this property when inserting or updating a
+      dataset in order to control who is allowed to access the data. If
+      unspecified at dataset creation time, BigQuery adds default dataset
+      access for the following entities: access.specialGroup: projectReaders;
+      access.role: READER; access.specialGroup: projectWriters; access.role:
+      WRITER; access.specialGroup: projectOwners; access.role: OWNER;
+      access.userByEmail: [dataset creator email]; access.role: OWNER;
+    creationTime: [Output-only] The time when this dataset was created, in
+      milliseconds since the epoch.
+    datasetReference: [Required] A reference that identifies the dataset.
+    defaultTableExpirationMs: [Optional] The default lifetime of all tables in
+      the dataset, in milliseconds. The minimum value is 3600000 milliseconds
+      (one hour). Once this property is set, all newly-created tables in the
+      dataset will have an expirationTime property set to the creation time
+      plus the value in this property, and changing the value will only affect
+      new tables, not existing ones. When the expirationTime for a given table
+      is reached, that table will be deleted automatically. If a table's
+      expirationTime is modified or removed before the table expires, or if
+      you provide an explicit expirationTime when creating a table, that value
+      takes precedence over the default expiration time indicated by this
+      property.
+    description: [Optional] A user-friendly description of the dataset.
+    etag: [Output-only] A hash of the resource.
+    friendlyName: [Optional] A descriptive name for the dataset.
+    id: [Output-only] The fully-qualified unique name of the dataset in the
+      format projectId:datasetId. The dataset name without the project name is
+      given in the datasetId field. When creating a new dataset, leave this
+      field blank, and instead specify the datasetId field.
+    kind: [Output-only] The resource type.
+    lastModifiedTime: [Output-only] The date when this dataset or any of its
+      tables was last modified, in milliseconds since the epoch.
+    location: [Experimental] The geographic location where the dataset should
+      reside. Possible values include EU and US. The default value is US.
+    selfLink: [Output-only] A URL that can be used to access the resource
+      again. You can use this URL in Get or Update requests to the resource.
+  """
+
+  class AccessValueListEntry(_messages.Message):
+    """A AccessValueListEntry object.
+
+    Fields:
+      domain: [Pick one] A domain to grant access to. Any users signed in with
+        the domain specified will be granted the specified access. Example:
+        "example.com".
+      groupByEmail: [Pick one] An email address of a Google Group to grant
+        access to.
+      role: [Required] Describes the rights granted to the user specified by
+        the other member of the access object. The following string values are
+        supported: READER, WRITER, OWNER.
+      specialGroup: [Pick one] A special group to grant access to. Possible
+        values include: projectOwners: Owners of the enclosing project.
+        projectReaders: Readers of the enclosing project. projectWriters:
+        Writers of the enclosing project. allAuthenticatedUsers: All
+        authenticated BigQuery users.
+      userByEmail: [Pick one] An email address of a user to grant access to.
+        For example: fred@example.com.
+      view: [Pick one] A view from a different dataset to grant access to.
+        Queries executed against that view will have read access to tables in
+        this dataset. The role field is not required when this field is set.
+        If that view is updated by any user, access to the view needs to be
+        granted again via an update operation.
+    """
+
+    domain = _messages.StringField(1)
+    groupByEmail = _messages.StringField(2)
+    role = _messages.StringField(3)
+    specialGroup = _messages.StringField(4)
+    userByEmail = _messages.StringField(5)
+    view = _messages.MessageField('TableReference', 6)
+
+  access = _messages.MessageField('AccessValueListEntry', 1, repeated=True)
+  creationTime = _messages.IntegerField(2)
+  datasetReference = _messages.MessageField('DatasetReference', 3)
+  defaultTableExpirationMs = _messages.IntegerField(4)
+  description = _messages.StringField(5)
+  etag = _messages.StringField(6)
+  friendlyName = _messages.StringField(7)
+  id = _messages.StringField(8)
+  kind = _messages.StringField(9, default=u'bigquery#dataset')
+  lastModifiedTime = _messages.IntegerField(10)
+  location = _messages.StringField(11)
+  selfLink = _messages.StringField(12)
+
+
+class DatasetList(_messages.Message):
+  """A DatasetList object.
+
+  Messages:
+    DatasetsValueListEntry: A DatasetsValueListEntry object.
+
+  Fields:
+    datasets: An array of the dataset resources in the project. Each resource
+      contains basic information. For full information about a particular
+      dataset resource, use the Datasets: get method. This property is omitted
+      when there are no datasets in the project.
+    etag: A hash value of the results page. You can use this property to
+      determine if the page has changed since the last request.
+    kind: The list type. This property always returns the value
+      "bigquery#datasetList".
+    nextPageToken: A token that can be used to request the next results page.
+      This property is omitted on the final results page.
+  """
+
+  class DatasetsValueListEntry(_messages.Message):
+    """A DatasetsValueListEntry object.
+
+    Fields:
+      datasetReference: The dataset reference. Use this property to access
+        specific parts of the dataset's ID, such as project ID or dataset ID.
+      friendlyName: A descriptive name for the dataset, if one exists.
+      id: The fully-qualified, unique, opaque ID of the dataset.
+      kind: The resource type. This property always returns the value
+        "bigquery#dataset".
+    """
+
+    datasetReference = _messages.MessageField('DatasetReference', 1)
+    friendlyName = _messages.StringField(2)
+    id = _messages.StringField(3)
+    kind = _messages.StringField(4, default=u'bigquery#dataset')
+
+  datasets = _messages.MessageField('DatasetsValueListEntry', 1, repeated=True)
+  etag = _messages.StringField(2)
+  kind = _messages.StringField(3, default=u'bigquery#datasetList')
+  nextPageToken = _messages.StringField(4)
+
+
+class DatasetReference(_messages.Message):
+  """A DatasetReference object.
+
+  Fields:
+    datasetId: [Required] A unique ID for this dataset, without the project
+      name. The ID must contain only letters (a-z, A-Z), numbers (0-9), or
+      underscores (_). The maximum length is 1,024 characters.
+    projectId: [Optional] The ID of the project containing this dataset.
+  """
+
+  datasetId = _messages.StringField(1)
+  projectId = _messages.StringField(2)
+
+
+class ErrorProto(_messages.Message):
+  """A ErrorProto object.
+
+  Fields:
+    debugInfo: Debugging information. This property is internal to Google and
+      should not be used.
+    location: Specifies where the error occurred, if present.
+    message: A human-readable description of the error.
+    reason: A short error code that summarizes the error.
+  """
+
+  debugInfo = _messages.StringField(1)
+  location = _messages.StringField(2)
+  message = _messages.StringField(3)
+  reason = _messages.StringField(4)
+
+
+class ExplainQueryStage(_messages.Message):
+  """A ExplainQueryStage object.
+
+  Fields:
+    computeRatioAvg: Relative amount of time the average shard spent on CPU-
+      bound tasks.
+    computeRatioMax: Relative amount of time the slowest shard spent on CPU-
+      bound tasks.
+    id: Unique ID for stage within plan.
+    name: Human-readable name for stage.
+    readRatioAvg: Relative amount of time the average shard spent reading
+      input.
+    readRatioMax: Relative amount of time the slowest shard spent reading
+      input.
+    recordsRead: Number of records read into the stage.
+    recordsWritten: Number of records written by the stage.
+    steps: List of operations within the stage in dependency order
+      (approximately chronological).
+    waitRatioAvg: Relative amount of time the average shard spent waiting to
+      be scheduled.
+    waitRatioMax: Relative amount of time the slowest shard spent waiting to
+      be scheduled.
+    writeRatioAvg: Relative amount of time the average shard spent on writing
+      output.
+    writeRatioMax: Relative amount of time the slowest shard spent on writing
+      output.
+  """
+
+  computeRatioAvg = _messages.FloatField(1)
+  computeRatioMax = _messages.FloatField(2)
+  id = _messages.IntegerField(3)
+  name = _messages.StringField(4)
+  readRatioAvg = _messages.FloatField(5)
+  readRatioMax = _messages.FloatField(6)
+  recordsRead = _messages.IntegerField(7)
+  recordsWritten = _messages.IntegerField(8)
+  steps = _messages.MessageField('ExplainQueryStep', 9, repeated=True)
+  waitRatioAvg = _messages.FloatField(10)
+  waitRatioMax = _messages.FloatField(11)
+  writeRatioAvg = _messages.FloatField(12)
+  writeRatioMax = _messages.FloatField(13)
+
+
+class ExplainQueryStep(_messages.Message):
+  """A ExplainQueryStep object.
+
+  Fields:
+    kind: Machine-readable operation type.
+    substeps: Human-readable stage descriptions.
+  """
+
+  kind = _messages.StringField(1)
+  substeps = _messages.StringField(2, repeated=True)
+
+
+class ExternalDataConfiguration(_messages.Message):
+  """A ExternalDataConfiguration object.
+
+  Fields:
+    autodetect: [Experimental] Try to detect schema and format options
+      automatically. Any option specified explicitly will be honored.
+    bigtableOptions: [Optional] Additional options if sourceFormat is set to
+      BIGTABLE.
+    compression: [Optional] The compression type of the data source. Possible
+      values include GZIP and NONE. The default value is NONE. This setting is
+      ignored for Google Cloud Bigtable, Google Cloud Datastore backups and
+      Avro formats.
+    csvOptions: Additional properties to set if sourceFormat is set to CSV.
+    ignoreUnknownValues: [Optional] Indicates if BigQuery should allow extra
+      values that are not represented in the table schema. If true, the extra
+      values are ignored. If false, records with extra columns are treated as
+      bad records, and if there are too many bad records, an invalid error is
+      returned in the job result. The default value is false. The sourceFormat
+      property determines what BigQuery treats as an extra value: CSV:
+      Trailing columns JSON: Named values that don't match any column names
+      Google Cloud Bigtable: This setting is ignored. Google Cloud Datastore
+      backups: This setting is ignored. Avro: This setting is ignored.
+    maxBadRecords: [Optional] The maximum number of bad records that BigQuery
+      can ignore when reading data. If the number of bad records exceeds this
+      value, an invalid error is returned in the job result. The default value
+      is 0, which requires that all records are valid. This setting is ignored
+      for Google Cloud Bigtable, Google Cloud Datastore backups and Avro
+      formats.
+    schema: [Optional] The schema for the data. Schema is required for CSV and
+      JSON formats. Schema is disallowed for Google Cloud Bigtable, Cloud
+      Datastore backups, and Avro formats.
+    sourceFormat: [Required] The data format. For CSV files, specify "CSV".
+      For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro
+      files, specify "AVRO". For Google Cloud Datastore backups, specify
+      "DATASTORE_BACKUP". [Experimental] For Google Cloud Bigtable, specify
+      "BIGTABLE". Please note that reading from Google Cloud Bigtable is
+      experimental and has to be enabled for your project. Please contact
+      Google Cloud Support to enable this for your project.
+    sourceUris: [Required] The fully-qualified URIs that point to your data in
+      Google Cloud. For Google Cloud Storage URIs: Each URI can contain one
+      '*' wildcard character and it must come after the 'bucket' name. Size
+      limits related to load jobs apply to external data sources, plus an
+      additional limit of 10 GB maximum size across all URIs. For Google Cloud
+      Bigtable URIs: Exactly one URI can be specified and it has be a fully
+      specified and valid HTTPS URL for a Google Cloud Bigtable table. For
+      Google Cloud Datastore backups, exactly one URI can be specified, and it
+      must end with '.backup_info'. Also, the '*' wildcard character is not
+      allowed.
+  """
+
+  autodetect = _messages.BooleanField(1)
+  bigtableOptions = _messages.MessageField('BigtableOptions', 2)
+  compression = _messages.StringField(3)
+  csvOptions = _messages.MessageField('CsvOptions', 4)
+  ignoreUnknownValues = _messages.BooleanField(5)
+  maxBadRecords = _messages.IntegerField(6, variant=_messages.Variant.INT32)
+  schema = _messages.MessageField('TableSchema', 7)
+  sourceFormat = _messages.StringField(8)
+  sourceUris = _messages.StringField(9, repeated=True)
+
+
+class GetQueryResultsResponse(_messages.Message):
+  """A GetQueryResultsResponse object.
+
+  Fields:
+    cacheHit: Whether the query result was fetched from the query cache.
+    errors: [Output-only] All errors and warnings encountered during the
+      running of the job. Errors here do not necessarily mean that the job has
+      completed or was unsuccessful.
+    etag: A hash of this response.
+    jobComplete: Whether the query has completed or not. If rows or totalRows
+      are present, this will always be true. If this is false, totalRows will
+      not be available.
+    jobReference: Reference to the BigQuery Job that was created to run the
+      query. This field will be present even if the original request timed
+      out, in which case GetQueryResults can be used to read the results once
+      the query has completed. Since this API only returns the first page of
+      results, subsequent pages can be fetched via the same mechanism
+      (GetQueryResults).
+    kind: The resource type of the response.
+    pageToken: A token used for paging results.
+    rows: An object with as many results as can be contained within the
+      maximum permitted reply size. To get any additional rows, you can call
+      GetQueryResults and specify the jobReference returned above. Present
+      only when the query completes successfully.
+    schema: The schema of the results. Present only when the query completes
+      successfully.
+    totalBytesProcessed: The total number of bytes processed for this query.
+    totalRows: The total number of rows in the complete query result set,
+      which can be more than the number of rows in this single page of
+      results. Present only when the query completes successfully.
+  """
+
+  cacheHit = _messages.BooleanField(1)
+  errors = _messages.MessageField('ErrorProto', 2, repeated=True)
+  etag = _messages.StringField(3)
+  jobComplete = _messages.BooleanField(4)
+  jobReference = _messages.MessageField('JobReference', 5)
+  kind = _messages.StringField(6, default=u'bigquery#getQueryResultsResponse')
+  pageToken = _messages.StringField(7)
+  rows = _messages.MessageField('TableRow', 8, repeated=True)
+  schema = _messages.MessageField('TableSchema', 9)
+  totalBytesProcessed = _messages.IntegerField(10)
+  totalRows = _messages.IntegerField(11, variant=_messages.Variant.UINT64)
+
+
+class IntervalPartitionConfiguration(_messages.Message):
+  """A IntervalPartitionConfiguration object.
+
+  Fields:
+    expirationMs: A string attribute.
+    type: A string attribute.
+  """
+
+  expirationMs = _messages.IntegerField(1)
+  type = _messages.StringField(2)
+
+
+class Job(_messages.Message):
+  """A Job object.
+
+  Fields:
+    configuration: [Required] Describes the job configuration.
+    etag: [Output-only] A hash of this resource.
+    id: [Output-only] Opaque ID field of the job
+    jobReference: [Optional] Reference describing the unique-per-user name of
+      the job.
+    kind: [Output-only] The type of the resource.
+    selfLink: [Output-only] A URL that can be used to access this resource
+      again.
+    statistics: [Output-only] Information about the job, including starting
+      time and ending time of the job.
+    status: [Output-only] The status of this job. Examine this value when
+      polling an asynchronous job to see if the job is complete.
+    user_email: [Output-only] Email address of the user who ran the job.
+  """
+
+  configuration = _messages.MessageField('JobConfiguration', 1)
+  etag = _messages.StringField(2)
+  id = _messages.StringField(3)
+  jobReference = _messages.MessageField('JobReference', 4)
+  kind = _messages.StringField(5, default=u'bigquery#job')
+  selfLink = _messages.StringField(6)
+  statistics = _messages.MessageField('JobStatistics', 7)
+  status = _messages.MessageField('JobStatus', 8)
+  user_email = _messages.StringField(9)
+
+
+class JobCancelResponse(_messages.Message):
+  """A JobCancelResponse object.
+
+  Fields:
+    job: The final state of the job.
+    kind: The resource type of the response.
+  """
+
+  job = _messages.MessageField('Job', 1)
+  kind = _messages.StringField(2, default=u'bigquery#jobCancelResponse')
+
+
+class JobConfiguration(_messages.Message):
+  """A JobConfiguration object.
+
+  Fields:
+    copy: [Pick one] Copies a table.
+    dryRun: [Optional] If set, don't actually run this job. A valid query will
+      return a mostly empty response with some processing statistics, while an
+      invalid query will return the same error it would if it wasn't a dry
+      run. Behavior of non-query jobs is undefined.
+    extract: [Pick one] Configures an extract job.
+    load: [Pick one] Configures a load job.
+    query: [Pick one] Configures a query job.
+  """
+
+  copy = _messages.MessageField('JobConfigurationTableCopy', 1)
+  dryRun = _messages.BooleanField(2)
+  extract = _messages.MessageField('JobConfigurationExtract', 3)
+  load = _messages.MessageField('JobConfigurationLoad', 4)
+  query = _messages.MessageField('JobConfigurationQuery', 5)
+
+
+class JobConfigurationExtract(_messages.Message):
+  """A JobConfigurationExtract object.
+
+  Fields:
+    compression: [Optional] The compression type to use for exported files.
+      Possible values include GZIP and NONE. The default value is NONE.
+    destinationFormat: [Optional] The exported file format. Possible values
+      include CSV, NEWLINE_DELIMITED_JSON and AVRO. The default value is CSV.
+      Tables with nested or repeated fields cannot be exported as CSV.
+    destinationUri: [Pick one] DEPRECATED: Use destinationUris instead,
+      passing only one URI as necessary. The fully-qualified Google Cloud
+      Storage URI where the extracted table should be written.
+    destinationUris: [Pick one] A list of fully-qualified Google Cloud Storage
+      URIs where the extracted table should be written.
+    fieldDelimiter: [Optional] Delimiter to use between fields in the exported
+      data. Default is ','
+    printHeader: [Optional] Whether to print out a header row in the results.
+      Default is true.
+    sourceTable: [Required] A reference to the table being exported.
+  """
+
+  compression = _messages.StringField(1)
+  destinationFormat = _messages.StringField(2)
+  destinationUri = _messages.StringField(3)
+  destinationUris = _messages.StringField(4, repeated=True)
+  fieldDelimiter = _messages.StringField(5)
+  printHeader = _messages.BooleanField(6, default=True)
+  sourceTable = _messages.MessageField('TableReference', 7)
+
+
+class JobConfigurationLoad(_messages.Message):
+  """A JobConfigurationLoad object.
+
+  Fields:
+    allowJaggedRows: [Optional] Accept rows that are missing trailing optional
+      columns. The missing values are treated as nulls. If false, records with
+      missing trailing columns are treated as bad records, and if there are
+      too many bad records, an invalid error is returned in the job result.
+      The default value is false. Only applicable to CSV, ignored for other
+      formats.
+    allowQuotedNewlines: Indicates if BigQuery should allow quoted data
+      sections that contain newline characters in a CSV file. The default
+      value is false.
+    createDisposition: [Optional] Specifies whether the job is allowed to
+      create new tables. The following values are supported: CREATE_IF_NEEDED:
+      If the table does not exist, BigQuery creates the table. CREATE_NEVER:
+      The table must already exist. If it does not, a 'notFound' error is
+      returned in the job result. The default value is CREATE_IF_NEEDED.
+      Creation, truncation and append actions occur as one atomic update upon
+      job completion.
+    destinationTable: [Required] The destination table to load the data into.
+    encoding: [Optional] The character encoding of the data. The supported
+      values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery
+      decodes the data after the raw, binary data has been split using the
+      values of the quote and fieldDelimiter properties.
+    fieldDelimiter: [Optional] The separator for fields in a CSV file. The
+      separator can be any ISO-8859-1 single-byte character. To use a
+      character in the range 128-255, you must encode the character as UTF8.
+      BigQuery converts the string to ISO-8859-1 encoding, and then uses the
+      first byte of the encoded string to split the data in its raw, binary
+      state. BigQuery also supports the escape sequence "\t" to specify a tab
+      separator. The default value is a comma (',').
+    ignoreUnknownValues: [Optional] Indicates if BigQuery should allow extra
+      values that are not represented in the table schema. If true, the extra
+      values are ignored. If false, records with extra columns are treated as
+      bad records, and if there are too many bad records, an invalid error is
+      returned in the job result. The default value is false. The sourceFormat
+      property determines what BigQuery treats as an extra value: CSV:
+      Trailing columns JSON: Named values that don't match any column names
+    maxBadRecords: [Optional] The maximum number of bad records that BigQuery
+      can ignore when running the job. If the number of bad records exceeds
+      this value, an invalid error is returned in the job result. The default
+      value is 0, which requires that all records are valid.
+    projectionFields: [Experimental] If sourceFormat is set to
+      "DATASTORE_BACKUP", indicates which entity properties to load into
+      BigQuery from a Cloud Datastore backup. Property names are case
+      sensitive and must be top-level properties. If no properties are
+      specified, BigQuery loads all properties. If any named property isn't
+      found in the Cloud Datastore backup, an invalid error is returned in the
+      job result.
+    quote: [Optional] The value that is used to quote data sections in a CSV
+      file. BigQuery converts the string to ISO-8859-1 encoding, and then uses
+      the first byte of the encoded string to split the data in its raw,
+      binary state. The default value is a double-quote ('"'). If your data
+      does not contain quoted sections, set the property value to an empty
+      string. If your data contains quoted newline characters, you must also
+      set the allowQuotedNewlines property to true.
+    schema: [Optional] The schema for the destination table. The schema can be
+      omitted if the destination table already exists, or if you're loading
+      data from Google Cloud Datastore.
+    schemaInline: [Deprecated] The inline schema. For CSV schemas, specify as
+      "Field1:Type1[,Field2:Type2]*". For example, "foo:STRING, bar:INTEGER,
+      baz:FLOAT".
+    schemaInlineFormat: [Deprecated] The format of the schemaInline property.
+    skipLeadingRows: [Optional] The number of rows at the top of a CSV file
+      that BigQuery will skip when loading the data. The default value is 0.
+      This property is useful if you have header rows in the file that should
+      be skipped.
+    sourceFormat: [Optional] The format of the data files. For CSV files,
+      specify "CSV". For datastore backups, specify "DATASTORE_BACKUP". For
+      newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro,
+      specify "AVRO". The default value is CSV.
+    sourceUris: [Required] The fully-qualified URIs that point to your data in
+      Google Cloud Storage. Each URI can contain one '*' wildcard character
+      and it must come after the 'bucket' name.
+    writeDisposition: [Optional] Specifies the action that occurs if the
+      destination table already exists. The following values are supported:
+      WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the
+      table data. WRITE_APPEND: If the table already exists, BigQuery appends
+      the data to the table. WRITE_EMPTY: If the table already exists and
+      contains data, a 'duplicate' error is returned in the job result. The
+      default value is WRITE_APPEND. Each action is atomic and only occurs if
+      BigQuery is able to complete the job successfully. Creation, truncation
+      and append actions occur as one atomic update upon job completion.
+  """
+
+  allowJaggedRows = _messages.BooleanField(1)
+  allowQuotedNewlines = _messages.BooleanField(2)
+  createDisposition = _messages.StringField(3)
+  destinationTable = _messages.MessageField('TableReference', 4)
+  encoding = _messages.StringField(5)
+  fieldDelimiter = _messages.StringField(6)
+  ignoreUnknownValues = _messages.BooleanField(7)
+  maxBadRecords = _messages.IntegerField(8, variant=_messages.Variant.INT32)
+  projectionFields = _messages.StringField(9, repeated=True)
+  quote = _messages.StringField(10, default=u'"')
+  schema = _messages.MessageField('TableSchema', 11)
+  schemaInline = _messages.StringField(12)
+  schemaInlineFormat = _messages.StringField(13)
+  skipLeadingRows = _messages.IntegerField(14, variant=_messages.Variant.INT32)
+  sourceFormat = _messages.StringField(15)
+  sourceUris = _messages.StringField(16, repeated=True)
+  writeDisposition = _messages.StringField(17)
+
+
+class JobConfigurationQuery(_messages.Message):
+  """A JobConfigurationQuery object.
+
+  Messages:
+    TableDefinitionsValue: [Optional] If querying an external data source
+      outside of BigQuery, describes the data format, location and other
+      properties of the data source. By defining these properties, the data
+      source can then be queried as if it were a standard BigQuery table.
+
+  Fields:
+    allowLargeResults: If true, allows the query to produce arbitrarily large
+      result tables at a slight cost in performance. Requires destinationTable
+      to be set.
+    createDisposition: [Optional] Specifies whether the job is allowed to
+      create new tables. The following values are supported: CREATE_IF_NEEDED:
+      If the table does not exist, BigQuery creates the table. CREATE_NEVER:
+      The table must already exist. If it does not, a 'notFound' error is
+      returned in the job result. The default value is CREATE_IF_NEEDED.
+      Creation, truncation and append actions occur as one atomic update upon
+      job completion.
+    defaultDataset: [Optional] Specifies the default dataset to use for
+      unqualified table names in the query.
+    destinationTable: [Optional] Describes the table where the query results
+      should be stored. If not present, a new table will be created to store
+      the results.
+    flattenResults: [Optional] Flattens all nested and repeated fields in the
+      query results. The default value is true. allowLargeResults must be true
+      if this is set to false.
+    maximumBillingTier: [Optional] Limits the billing tier for this job.
+      Queries that have resource usage beyond this tier will fail (without
+      incurring a charge). If unspecified, this will be set to your project
+      default.
+    preserveNulls: [Deprecated] This property is deprecated.
+    priority: [Optional] Specifies a priority for the query. Possible values
+      include INTERACTIVE and BATCH. The default value is INTERACTIVE.
+    query: [Required] BigQuery SQL query to execute.
+    tableDefinitions: [Optional] If querying an external data source outside
+      of BigQuery, describes the data format, location and other properties of
+      the data source. By defining these properties, the data source can then
+      be queried as if it were a standard BigQuery table.
+    useLegacySql: [Experimental] Specifies whether to use BigQuery's legacy
+      SQL dialect for this query. The default value is true. If set to false,
+      the query will use BigQuery's updated SQL dialect with improved
+      standards compliance. When using BigQuery's updated SQL, the values of
+      allowLargeResults and flattenResults are ignored. Queries with
+      useLegacySql set to false will be run as if allowLargeResults is true
+      and flattenResults is false.
+    useQueryCache: [Optional] Whether to look for the result in the query
+      cache. The query cache is a best-effort cache that will be flushed
+      whenever tables in the query are modified. Moreover, the query cache is
+      only available when a query does not have a destination table specified.
+      The default value is true.
+    userDefinedFunctionResources: [Experimental] Describes user-defined
+      function resources used in the query.
+    writeDisposition: [Optional] Specifies the action that occurs if the
+      destination table already exists. The following values are supported:
+      WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the
+      table data. WRITE_APPEND: If the table already exists, BigQuery appends
+      the data to the table. WRITE_EMPTY: If the table already exists and
+      contains data, a 'duplicate' error is returned in the job result. The
+      default value is WRITE_EMPTY. Each action is atomic and only occurs if
+      BigQuery is able to complete the job successfully. Creation, truncation
+      and append actions occur as one atomic update upon job completion.
+  """
+
+  @encoding.MapUnrecognizedFields('additionalProperties')
+  class TableDefinitionsValue(_messages.Message):
+    """[Optional] If querying an external data source outside of BigQuery,
+    describes the data format, location and other properties of the data
+    source. By defining these properties, the data source can then be queried
+    as if it were a standard BigQuery table.
+
+    Messages:
+      AdditionalProperty: An additional property for a TableDefinitionsValue
+        object.
+
+    Fields:
+      additionalProperties: Additional properties of type
+        TableDefinitionsValue
+    """
+
+    class AdditionalProperty(_messages.Message):
+      """An additional property for a TableDefinitionsValue object.
+
+      Fields:
+        key: Name of the additional property.
+        value: A ExternalDataConfiguration attribute.
+      """
+
+      key = _messages.StringField(1)
+      value = _messages.MessageField('ExternalDataConfiguration', 2)
+
+    additionalProperties = _messages.MessageField('AdditionalProperty', 1, repeated=True)
+
+  allowLargeResults = _messages.BooleanField(1)
+  createDisposition = _messages.StringField(2)
+  defaultDataset = _messages.MessageField('DatasetReference', 3)
+  destinationTable = _messages.MessageField('TableReference', 4)
+  flattenResults = _messages.BooleanField(5, default=True)
+  maximumBillingTier = _messages.IntegerField(6, variant=_messages.Variant.INT32, default=1)
+  preserveNulls = _messages.BooleanField(7)
+  priority = _messages.StringField(8)
+  query = _messages.StringField(9)
+  tableDefinitions = _messages.MessageField('TableDefinitionsValue', 10)
+  useLegacySql = _messages.BooleanField(11)
+  useQueryCache = _messages.BooleanField(12, default=True)
+  userDefinedFunctionResources = _messages.MessageField('UserDefinedFunctionResource', 13, repeated=True)
+  writeDisposition = _messages.StringField(14)
+
+
+class JobConfigurationTableCopy(_messages.Message):
+  """A JobConfigurationTableCopy object.
+
+  Fields:
+    createDisposition: [Optional] Specifies whether the job is allowed to
+      create new tables. The following values are supported: CREATE_IF_NEEDED:
+      If the table does not exist, BigQuery creates the table. CREATE_NEVER:
+      The table must already exist. If it does not, a 'notFound' error is
+      returned in the job result. The default value is CREATE_IF_NEEDED.
+      Creation, truncation and append actions occur as one atomic update upon
+      job completion.
+    destinationTable: [Required] The destination table
+    sourceTable: [Pick one] Source table to copy.
+    sourceTables: [Pick one] Source tables to copy.
+    writeDisposition: [Optional] Specifies the action that occurs if the
+      destination table already exists. The following values are supported:
+      WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the
+      table data. WRITE_APPEND: If the table already exists, BigQuery appends
+      the data to the table. WRITE_EMPTY: If the table already exists and
+      contains data, a 'duplicate' error is returned in the job result. The
+      default value is WRITE_EMPTY. Each action is atomic and only occurs if
+      BigQuery is able to complete the job successfully. Creation, truncation
+      and append actions occur as one atomic update upon job completion.
+  """
+
+  createDisposition = _messages.StringField(1)
+  destinationTable = _messages.MessageField('TableReference', 2)
+  sourceTable = _messages.MessageField('TableReference', 3)
+  sourceTables = _messages.MessageField('TableReference', 4, repeated=True)
+  writeDisposition = _messages.StringField(5)
+
+
+class JobList(_messages.Message):
+  """A JobList object.
+
+  Messages:
+    JobsValueListEntry: A JobsValueListEntry object.
+
+  Fields:
+    etag: A hash of this page of results.
+    jobs: List of jobs that were requested.
+    kind: The resource type of the response.
+    nextPageToken: A token to request the next page of results.
+  """
+
+  class JobsValueListEntry(_messages.Message):
+    """A JobsValueListEntry object.
+
+    Fields:
+      configuration: [Full-projection-only] Specifies the job configuration.
+      errorResult: A result object that will be present only if the job has
+        failed.
+      id: Unique opaque ID of the job.
+      jobReference: Job reference uniquely identifying the job.
+      kind: The resource type.
+      state: Running state of the job. When the state is DONE, errorResult can
+        be checked to determine whether the job succeeded or failed.
+      statistics: [Output-only] Information about the job, including starting
+        time and ending time of the job.
+      status: [Full-projection-only] Describes the state of the job.
+      user_email: [Full-projection-only] Email address of the user who ran the
+        job.
+    """
+
+    configuration = _messages.MessageField('JobConfiguration', 1)
+    errorResult = _messages.MessageField('ErrorProto', 2)
+    id = _messages.StringField(3)
+    jobReference = _messages.MessageField('JobReference', 4)
+    kind = _messages.StringField(5, default=u'bigquery#job')
+    state = _messages.StringField(6)
+    statistics = _messages.MessageField('JobStatistics', 7)
+    status = _messages.MessageField('JobStatus', 8)
+    user_email = _messages.StringField(9)
+
+  etag = _messages.StringField(1)
+  jobs = _messages.MessageField('JobsValueListEntry', 2, repeated=True)
+  kind = _messages.StringField(3, default=u'bigquery#jobList')
+  nextPageToken = _messages.StringField(4)
+
+
+class JobReference(_messages.Message):
+  """A JobReference object.
+
+  Fields:
+    jobId: [Required] The ID of the job. The ID must contain only letters
+      (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum
+      length is 1,024 characters.
+    projectId: [Required] The ID of the project containing this job.
+  """
+
+  jobId = _messages.StringField(1)
+  projectId = _messages.StringField(2)
+
+
+class JobStatistics(_messages.Message):
+  """A JobStatistics object.
+
+  Fields:
+    creationTime: [Output-only] Creation time of this job, in milliseconds
+      since the epoch. This field will be present on all jobs.
+    endTime: [Output-only] End time of this job, in milliseconds since the
+      epoch. This field will be present whenever a job is in the DONE state.
+    extract: [Output-only] Statistics for an extract job.
+    load: [Output-only] Statistics for a load job.
+    query: [Output-only] Statistics for a query job.
+    startTime: [Output-only] Start time of this job, in milliseconds since the
+      epoch. This field will be present when the job transitions from the
+      PENDING state to either RUNNING or DONE.
+    totalBytesProcessed: [Output-only] [Deprecated] Use the bytes processed in
+      the query statistics instead.
+  """
+
+  creationTime = _messages.IntegerField(1)
+  endTime = _messages.IntegerField(2)
+  extract = _messages.MessageField('JobStatistics4', 3)
+  load = _messages.MessageField('JobStatistics3', 4)
+  query = _messages.MessageField('JobStatistics2', 5)
+  startTime = _messages.IntegerField(6)
+  totalBytesProcessed = _messages.IntegerField(7)
+
+
+class JobStatistics2(_messages.Message):
+  """A JobStatistics2 object.
+
+  Fields:
+    billingTier: [Output-only] Billing tier for the job.
+    cacheHit: [Output-only] Whether the query result was fetched from the
+      query cache.
+    queryPlan: [Output-only, Experimental] Describes execution plan for the
+      query as a list of stages.
+    referencedTables: [Output-only, Experimental] Referenced tables for the
+      job. Queries that reference more than 50 tables will not have a complete
+      list.
+    totalBytesBilled: [Output-only] Total bytes billed for the job.
+    totalBytesProcessed: [Output-only] Total bytes processed for the job.
+  """
+
+  billingTier = _messages.IntegerField(1, variant=_messages.Variant.INT32)
+  cacheHit = _messages.BooleanField(2)
+  queryPlan = _messages.MessageField('ExplainQueryStage', 3, repeated=True)
+  referencedTables = _messages.MessageField('TableReference', 4, repeated=True)
+  totalBytesBilled = _messages.IntegerField(5)
+  totalBytesProcessed = _messages.IntegerField(6)
+
+
+class JobStatistics3(_messages.Message):
+  """A JobStatistics3 object.
+
+  Fields:
+    inputFileBytes: [Output-only] Number of bytes of source data in a load
+      job.
+    inputFiles: [Output-only] Number of source files in a load job.
+    outputBytes: [Output-only] Size of the loaded data in bytes. Note that
+      while a load job is in the running state, this value may change.
+    outputRows: [Output-only] Number of rows imported in a load job. Note that
+      while an import job is in the running state, this value may change.
+  """
+
+  inputFileBytes = _messages.IntegerField(1)
+  inputFiles = _messages.IntegerField(2)
+  outputBytes = _messages.IntegerField(3)
+  outputRows = _messages.IntegerField(4)
+
+
+class JobStatistics4(_messages.Message):
+  """A JobStatistics4 object.
+
+  Fields:
+    destinationUriFileCounts: [Output-only] Number of files per destination
+      URI or URI pattern specified in the extract configuration. These values
+      will be in the same order as the URIs specified in the 'destinationUris'
+      field.
+  """
+
+  destinationUriFileCounts = _messages.IntegerField(1, repeated=True)
+
+
+class JobStatus(_messages.Message):
+  """A JobStatus object.
+
+  Fields:
+    errorResult: [Output-only] Final error result of the job. If present,
+      indicates that the job has completed and was unsuccessful.
+    errors: [Output-only] All errors encountered during the running of the
+      job. Errors here do not necessarily mean that the job has completed or
+      was unsuccessful.
+    state: [Output-only] Running state of the job.
+  """
+
+  errorResult = _messages.MessageField('ErrorProto', 1)
+  errors = _messages.MessageField('ErrorProto', 2, repeated=True)
+  state = _messages.StringField(3)
+
+
+@encoding.MapUnrecognizedFields('additionalProperties')
+class JsonObject(_messages.Message):
+  """Represents a single JSON object.
+
+  Messages:
+    AdditionalProperty: An additional property for a JsonObject object.
+
+  Fields:
+    additionalProperties: Additional properties of type JsonObject
+  """
+
+  class AdditionalProperty(_messages.Message):
+    """An additional property for a JsonObject object.
+
+    Fields:
+      key: Name of the additional property.
+      value: A JsonValue attribute.
+    """
+
+    key = _messages.StringField(1)
+    value = _messages.MessageField('JsonValue', 2)
+
+  additionalProperties = _messages.MessageField('AdditionalProperty', 1, repeated=True)
+
+
+JsonValue = extra_types.JsonValue
+
+
+class ProjectList(_messages.Message):
+  """A ProjectList object.
+
+  Messages:
+    ProjectsValueListEntry: A ProjectsValueListEntry object.
+
+  Fields:
+    etag: A hash of the page of results
+    kind: The type of list.
+    nextPageToken: A token to request the next page of results.
+    projects: Projects to which you have at least READ access.
+    totalItems: The total number of projects in the list.
+  """
+
+  class ProjectsValueListEntry(_messages.Message):
+    """A ProjectsValueListEntry object.
+
+    Fields:
+      friendlyName: A descriptive name for this project.
+      id: An opaque ID of this project.
+      kind: The resource type.
+      numericId: The numeric ID of this project.
+      projectReference: A unique reference to this project.
+    """
+
+    friendlyName = _messages.StringField(1)
+    id = _messages.StringField(2)
+    kind = _messages.StringField(3, default=u'bigquery#project')
+    numericId = _messages.IntegerField(4, variant=_messages.Variant.UINT64)
+    projectReference = _messages.MessageField('ProjectReference', 5)
+
+  etag = _messages.StringField(1)
+  kind = _messages.StringField(2, default=u'bigquery#projectList')
+  nextPageToken = _messages.StringField(3)
+  projects = _messages.MessageField('ProjectsValueListEntry', 4, repeated=True)
+  totalItems = _messages.IntegerField(5, variant=_messages.Variant.INT32)
+
+
+class ProjectReference(_messages.Message):
+  """A ProjectReference object.
+
+  Fields:
+    projectId: [Required] ID of the project. Can be either the numeric ID or
+      the assigned ID of the project.
+  """
+
+  projectId = _messages.StringField(1)
+
+
+class QueryRequest(_messages.Message):
+  """A QueryRequest object.
+
+  Fields:
+    defaultDataset: [Optional] Specifies the default datasetId and projectId
+      to assume for any unqualified table names in the query. If not set, all
+      table names in the query string must be qualified in the format
+      'datasetId.tableId'.
+    dryRun: [Optional] If set to true, BigQuery doesn't run the job. Instead,
+      if the query is valid, BigQuery returns statistics about the job such as
+      how many bytes would be processed. If the query is invalid, an error
+      returns. The default value is false.
+    kind: The resource type of the request.
+    maxResults: [Optional] The maximum number of rows of data to return per
+      page of results. Setting this flag to a small value such as 1000 and
+      then paging through results might improve reliability when the query
+      result set is large. In addition to this limit, responses are also
+      limited to 10 MB. By default, there is no maximum row count, and only
+      the byte limit applies.
+    preserveNulls: [Deprecated] This property is deprecated.
+    query: [Required] A query string, following the BigQuery query syntax, of
+      the query to execute. Example: "SELECT count(f1) FROM
+      [myProjectId:myDatasetId.myTableId]".
+    timeoutMs: [Optional] How long to wait for the query to complete, in
+      milliseconds, before the request times out and returns. Note that this
+      is only a timeout for the request, not the query. If the query takes
+      longer to run than the timeout value, the call returns without any
+      results and with the 'jobComplete' flag set to false. You can call
+      GetQueryResults() to wait for the query to complete and read the
+      results. The default value is 10000 milliseconds (10 seconds).
+    useLegacySql: [Experimental] Specifies whether to use BigQuery's legacy
+      SQL dialect for this query. The default value is true. If set to false,
+      the query will use BigQuery's updated SQL dialect with improved
+      standards compliance. When using BigQuery's updated SQL, the values of
+      allowLargeResults and flattenResults are ignored. Queries with
+      useLegacySql set to false will be run as if allowLargeResults is true
+      and flattenResults is false.
+    useQueryCache: [Optional] Whether to look for the result in the query
+      cache. The query cache is a best-effort cache that will be flushed
+      whenever tables in the query are modified. The default value is true.
+  """
+
+  defaultDataset = _messages.MessageField('DatasetReference', 1)
+  dryRun = _messages.BooleanField(2)
+  kind = _messages.StringField(3, default=u'bigquery#queryRequest')
+  maxResults = _messages.IntegerField(4, variant=_messages.Variant.UINT32)
+  preserveNulls = _messages.BooleanField(5)
+  query = _messages.StringField(6)
+  timeoutMs = _messages.IntegerField(7, variant=_messages.Variant.UINT32)
+  useLegacySql = _messages.BooleanField(8)
+  useQueryCache = _messages.BooleanField(9, default=True)
+
+
+class QueryResponse(_messages.Message):
+  """A QueryResponse object.
+
+  Fields:
+    cacheHit: Whether the query result was fetched from the query cache.
+    errors: [Output-only] All errors and warnings encountered during the
+      running of the job. Errors here do not necessarily mean that the job has
+      completed or was unsuccessful.
+    jobComplete: Whether the query has completed or not. If rows or totalRows
+      are present, this will always be true. If this is false, totalRows will
+      not be available.
+    jobReference: Reference to the Job that was created to run the query. This
+      field will be present even if the original request timed out, in which
+      case GetQueryResults can be used to read the results once the query has
+      completed. Since this API only returns the first page of results,
+      subsequent pages can be fetched via the same mechanism
+      (GetQueryResults).
+    kind: The resource type.
+    pageToken: A token used for paging results.
+    rows: An object with as many results as can be contained within the
+      maximum permitted reply size. To get any additional rows, you can call
+      GetQueryResults and specify the jobReference returned above.
+    schema: The schema of the results. Present only when the query completes
+      successfully.
+    totalBytesProcessed: The total number of bytes processed for this query.
+      If this query was a dry run, this is the number of bytes that would be
+      processed if the query were run.
+    totalRows: The total number of rows in the complete query result set,
+      which can be more than the number of rows in this single page of
+      results.
+  """
+
+  cacheHit = _messages.BooleanField(1)
+  errors = _messages.MessageField('ErrorProto', 2, repeated=True)
+  jobComplete = _messages.BooleanField(3)
+  jobReference = _messages.MessageField('JobReference', 4)
+  kind = _messages.StringField(5, default=u'bigquery#queryResponse')
+  pageToken = _messages.StringField(6)
+  rows = _messages.MessageField('TableRow', 7, repeated=True)
+  schema = _messages.MessageField('TableSchema', 8)
+  totalBytesProcessed = _messages.IntegerField(9)
+  totalRows = _messages.IntegerField(10, variant=_messages.Variant.UINT64)
+
+
+class StandardQueryParameters(_messages.Message):
+  """Query parameters accepted by all methods.
+
+  Enums:
+    AltValueValuesEnum: Data format for the response.
+
+  Fields:
+    alt: Data format for the response.
+    fields: Selector specifying which fields to include in a partial response.
+    key: API key. Your API key identifies your project and provides you with
+      API access, quota, and reports. Required unless you provide an OAuth 2.0
+      token.
+    oauth_token: OAuth 2.0 token for the current user.
+    prettyPrint: Returns response with indentations and line breaks.
+    quotaUser: Available to use for quota purposes for server-side
+      applications. Can be any arbitrary string assigned to a user, but should
+      not exceed 40 characters. Overrides userIp if both are provided.
+    trace: A tracing token of the form "token:<tokenid>" to include in api
+      requests.
+    userIp: IP address of the site where the request originates. Use this if
+      you want to enforce per-user limits.
+  """
+
+  class AltValueValuesEnum(_messages.Enum):
+    """Data format for the response.
+
+    Values:
+      csv: Responses with Content-Type of text/csv
+      json: Responses with Content-Type of application/json
+    """
+    csv = 0
+    json = 1
+
+  alt = _messages.EnumField('AltValueValuesEnum', 1, default=u'json')
+  fields = _messages.StringField(2)
+  key = _messages.StringField(3)
+  oauth_token = _messages.StringField(4)
+  prettyPrint = _messages.BooleanField(5, default=True)
+  quotaUser = _messages.StringField(6)
+  trace = _messages.StringField(7)
+  userIp = _messages.StringField(8)
+
+
+class Streamingbuffer(_messages.Message):
+  """A Streamingbuffer object.
+
+  Fields:
+    estimatedBytes: [Output-only] A lower-bound estimate of the number of
+      bytes currently in the streaming buffer.
+    estimatedRows: [Output-only] A lower-bound estimate of the number of rows
+      currently in the streaming buffer.
+    oldestEntryTime: [Output-only] Contains the timestamp of the oldest entry
+      in the streaming buffer, in milliseconds since the epoch, if the
+      streaming buffer is available.
+  """
+
+  estimatedBytes = _messages.IntegerField(1, variant=_messages.Variant.UINT64)
+  estimatedRows = _messages.IntegerField(2, variant=_messages.Variant.UINT64)
+  oldestEntryTime = _messages.IntegerField(3, variant=_messages.Variant.UINT64)
+
+
+class Table(_messages.Message):
+  """A Table object.
+
+  Fields:
+    creationTime: [Output-only] The time when this table was created, in
+      milliseconds since the epoch.
+    description: [Optional] A user-friendly description of this table.
+    etag: [Output-only] A hash of this resource.
+    expirationTime: [Optional] The time when this table expires, in
+      milliseconds since the epoch. If not present, the table will persist
+      indefinitely. Expired tables will be deleted and their storage
+      reclaimed.
+    externalDataConfiguration: [Optional] Describes the data format, location,
+      and other properties of a table stored outside of BigQuery. By defining
+      these properties, the data source can then be queried as if it were a
+      standard BigQuery table.
+    friendlyName: [Optional] A descriptive name for this table.
+    id: [Output-only] An opaque ID uniquely identifying the table.
+    kind: [Output-only] The type of the resource.
+    lastModifiedTime: [Output-only] The time when this table was last
+      modified, in milliseconds since the epoch.
+    location: [Output-only] The geographic location where the table resides.
+      This value is inherited from the dataset.
+    numBytes: [Output-only] The size of this table in bytes, excluding any
+      data in the streaming buffer.
+    numRows: [Output-only] The number of rows of data in this table, excluding
+      any data in the streaming buffer.
+    partitionConfigurations: [Experimental] List of partition configurations
+      for this table. Currently only one configuration can be specified and it
+      can only be an interval partition with type daily.
+    schema: [Optional] Describes the schema of this table.
+    selfLink: [Output-only] A URL that can be used to access this resource
+      again.
+    streamingBuffer: [Output-only] Contains information regarding this table's
+      streaming buffer, if one is present. This field will be absent if the
+      table is not being streamed to or if there is no data in the streaming
+      buffer.
+    tableReference: [Required] Reference describing the ID of this table.
+    type: [Output-only] Describes the table type. The following values are
+      supported: TABLE: A normal BigQuery table. VIEW: A virtual table defined
+      by a SQL query. EXTERNAL: A table that references data stored in an
+      external storage system, such as Google Cloud Storage. The default value
+      is TABLE.
+    view: [Optional] The view definition.
+  """
+
+  creationTime = _messages.IntegerField(1)
+  description = _messages.StringField(2)
+  etag = _messages.StringField(3)
+  expirationTime = _messages.IntegerField(4)
+  externalDataConfiguration = _messages.MessageField('ExternalDataConfiguration', 5)
+  friendlyName = _messages.StringField(6)
+  id = _messages.StringField(7)
+  kind = _messages.StringField(8, default=u'bigquery#table')
+  lastModifiedTime = _messages.IntegerField(9, variant=_messages.Variant.UINT64)
+  location = _messages.StringField(10)
+  numBytes = _messages.IntegerField(11)
+  numRows = _messages.IntegerField(12, variant=_messages.Variant.UINT64)
+  partitionConfigurations = _messages.MessageField('TablePartitionConfiguration', 13, repeated=True)
+  schema = _messages.MessageField('TableSchema', 14)
+  selfLink = _messages.StringField(15)
+  streamingBuffer = _messages.MessageField('Streamingbuffer', 16)
+  tableReference = _messages.MessageField('TableReference', 17)
+  type = _messages.StringField(18)
+  view = _messages.MessageField('ViewDefinition', 19)
+
+
+class TableCell(_messages.Message):
+  """A TableCell object.
+
+  Fields:
+    v: A extra_types.JsonValue attribute.
+  """
+
+  v = _messages.MessageField('extra_types.JsonValue', 1)
+
+
+class TableDataInsertAllRequest(_messages.Message):
+  """A TableDataInsertAllRequest object.
+
+  Messages:
+    RowsValueListEntry: A RowsValueListEntry object.
+
+  Fields:
+    ignoreUnknownValues: [Optional] Accept rows that contain values that do
+      not match the schema. The unknown values are ignored. Default is false,
+      which treats unknown values as errors.
+    kind: The resource type of the response.
+    rows: The rows to insert.
+    skipInvalidRows: [Optional] Insert all valid rows of a request, even if
+      invalid rows exist. The default value is false, which causes the entire
+      request to fail if any invalid rows exist.
+    templateSuffix: [Experimental] If specified, treats the destination table
+      as a base template, and inserts the rows into an instance table named
+      "{destination}{templateSuffix}". BigQuery will manage creation of the
+      instance table, using the schema of the base template table. See
+      https://cloud.google.com/bigquery/streaming-data-into-bigquery#template-
+      tables for considerations when working with templates tables.
+  """
+
+  class RowsValueListEntry(_messages.Message):
+    """A RowsValueListEntry object.
+
+    Fields:
+      insertId: [Optional] A unique ID for each row. BigQuery uses this
+        property to detect duplicate insertion requests on a best-effort
+        basis.
+      json: [Required] A JSON object that contains a row of data. The object's
+        properties and values must match the destination table's schema.
+    """
+
+    insertId = _messages.StringField(1)
+    json = _messages.MessageField('JsonObject', 2)
+
+  ignoreUnknownValues = _messages.BooleanField(1)
+  kind = _messages.StringField(2, default=u'bigquery#tableDataInsertAllRequest')
+  rows = _messages.MessageField('RowsValueListEntry', 3, repeated=True)
+  skipInvalidRows = _messages.BooleanField(4)
+  templateSuffix = _messages.StringField(5)
+
+
+class TableDataInsertAllResponse(_messages.Message):
+  """A TableDataInsertAllResponse object.
+
+  Messages:
+    InsertErrorsValueListEntry: A InsertErrorsValueListEntry object.
+
+  Fields:
+    insertErrors: An array of errors for rows that were not inserted.
+    kind: The resource type of the response.
+  """
+
+  class InsertErrorsValueListEntry(_messages.Message):
+    """A InsertErrorsValueListEntry object.
+
+    Fields:
+      errors: Error information for the row indicated by the index property.
+      index: The index of the row that error applies to.
+    """
+
+    errors = _messages.MessageField('ErrorProto', 1, repeated=True)
+    index = _messages.IntegerField(2, variant=_messages.Variant.UINT32)
+
+  insertErrors = _messages.MessageField('InsertErrorsValueListEntry', 1, repeated=True)
+  kind = _messages.StringField(2, default=u'bigquery#tableDataInsertAllResponse')
+
+
+class TableDataList(_messages.Message):
+  """A TableDataList object.
+
+  Fields:
+    etag: A hash of this page of results.
+    kind: The resource type of the response.
+    pageToken: A token used for paging results. Providing this token instead
+      of the startIndex parameter can help you retrieve stable results when an
+      underlying table is changing.
+    rows: Rows of results.
+    totalRows: The total number of rows in the complete table.
+  """
+
+  etag = _messages.StringField(1)
+  kind = _messages.StringField(2, default=u'bigquery#tableDataList')
+  pageToken = _messages.StringField(3)
+  rows = _messages.MessageField('TableRow', 4, repeated=True)
+  totalRows = _messages.IntegerField(5)
+
+
+class TableFieldSchema(_messages.Message):
+  """A TableFieldSchema object.
+
+  Fields:
+    description: [Optional] The field description. The maximum length is 16K
+      characters.
+    fields: [Optional] Describes the nested schema fields if the type property
+      is set to RECORD.
+    mode: [Optional] The field mode. Possible values include NULLABLE,
+      REQUIRED and REPEATED. The default value is NULLABLE.
+    name: [Required] The field name. The name must contain only letters (a-z,
+      A-Z), numbers (0-9), or underscores (_), and must start with a letter or
+      underscore. The maximum length is 128 characters.
+    type: [Required] The field data type. Possible values include STRING,
+      BYTES, INTEGER, FLOAT, BOOLEAN, TIMESTAMP or RECORD (where RECORD
+      indicates that the field contains a nested schema).
+  """
+
+  description = _messages.StringField(1)
+  fields = _messages.MessageField('TableFieldSchema', 2, repeated=True)
+  mode = _messages.StringField(3)
+  name = _messages.StringField(4)
+  type = _messages.StringField(5)
+
+
+class TableList(_messages.Message):
+  """A TableList object.
+
+  Messages:
+    TablesValueListEntry: A TablesValueListEntry object.
+
+  Fields:
+    etag: A hash of this page of results.
+    kind: The type of list.
+    nextPageToken: A token to request the next page of results.
+    tables: Tables in the requested dataset.
+    totalItems: The total number of tables in the dataset.
+  """
+
+  class TablesValueListEntry(_messages.Message):
+    """A TablesValueListEntry object.
+
+    Fields:
+      friendlyName: The user-friendly name for this table.
+      id: An opaque ID of the table
+      kind: The resource type.
+      tableReference: A reference uniquely identifying the table.
+      type: The type of table. Possible values are: TABLE, VIEW.
+    """
+
+    friendlyName = _messages.StringField(1)
+    id = _messages.StringField(2)
+    kind = _messages.StringField(3, default=u'bigquery#table')
+    tableReference = _messages.MessageField('TableReference', 4)
+    type = _messages.StringField(5)
+
+  etag = _messages.StringField(1)
+  kind = _messages.StringField(2, default=u'bigquery#tableList')
+  nextPageToken = _messages.StringField(3)
+  tables = _messages.MessageField('TablesValueListEntry', 4, repeated=True)
+  totalItems = _messages.IntegerField(5, variant=_messages.Variant.INT32)
+
+
+class TablePartitionConfiguration(_messages.Message):
+  """[Required] A partition configuration. Only one type of partition should
+  be configured.
+
+  Fields:
+    interval: [Pick one] Configures an interval partition.
+  """
+
+  interval = _messages.MessageField('IntervalPartitionConfiguration', 1)
+
+
+class TableReference(_messages.Message):
+  """A TableReference object.
+
+  Fields:
+    datasetId: [Required] The ID of the dataset containing this table.
+    projectId: [Required] The ID of the project containing this table.
+    tableId: [Required] The ID of the table. The ID must contain only letters
+      (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is
+      1,024 characters.
+  """
+
+  datasetId = _messages.StringField(1)
+  projectId = _messages.StringField(2)
+  tableId = _messages.StringField(3)
+
+
+class TableRow(_messages.Message):
+  """A TableRow object.
+
+  Fields:
+    f: Represents a single row in the result set, consisting of one or more
+      fields.
+  """
+
+  f = _messages.MessageField('TableCell', 1, repeated=True)
+
+
+class TableSchema(_messages.Message):
+  """A TableSchema object.
+
+  Fields:
+    fields: Describes the fields in a table.
+  """
+
+  fields = _messages.MessageField('TableFieldSchema', 1, repeated=True)
+
+
+class UserDefinedFunctionResource(_messages.Message):
+  """A UserDefinedFunctionResource object.
+
+  Fields:
+    inlineCode: [Pick one] An inline resource that contains code for a user-
+      defined function (UDF). Providing a inline code resource is equivalent
+      to providing a URI for a file containing the same code.
+    resourceUri: [Pick one] A code resource to load from a Google Cloud
+      Storage URI (gs://bucket/path).
+  """
+
+  inlineCode = _messages.StringField(1)
+  resourceUri = _messages.StringField(2)
+
+
+class ViewDefinition(_messages.Message):
+  """A ViewDefinition object.
+
+  Fields:
+    query: [Required] A query that BigQuery executes when the view is
+      referenced.
+    userDefinedFunctionResources: [Experimental] Describes user-defined
+      function resources used in the query.
+  """
+
+  query = _messages.StringField(1)
+  userDefinedFunctionResources = _messages.MessageField('UserDefinedFunctionResource', 2, repeated=True)
+
+

http://git-wip-us.apache.org/repos/asf/beam/blob/59ad58ac/sdks/python/apache_beam/io/gcp/internal/clients/storage/__init__.py
----------------------------------------------------------------------
diff --git a/sdks/python/apache_beam/io/gcp/internal/clients/storage/__init__.py b/sdks/python/apache_beam/io/gcp/internal/clients/storage/__init__.py
new file mode 100644
index 0000000..8a726ef
--- /dev/null
+++ b/sdks/python/apache_beam/io/gcp/internal/clients/storage/__init__.py
@@ -0,0 +1,33 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements.  See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License.  You may obtain a copy of the License at
+#
+#    http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+#
+
+"""Common imports for generated storage client library."""
+# pylint:disable=wildcard-import
+
+import pkgutil
+
+# Protect against environments where apitools library is not available.
+# pylint: disable=wrong-import-order, wrong-import-position
+try:
+  from apitools.base.py import *
+  from apache_beam.io.gcp.internal.clients.storage.storage_v1_client import *
+  from apache_beam.io.gcp.internal.clients.storage.storage_v1_messages import *
+except ImportError:
+  pass
+# pylint: enable=wrong-import-order, wrong-import-position
+
+__path__ = pkgutil.extend_path(__path__, __name__)


Mime
View raw message