- 3.25.0 (latest)
- 3.24.0
- 3.23.1
- 3.22.0
- 3.21.0
- 3.20.1
- 3.19.0
- 3.18.0
- 3.17.2
- 3.16.0
- 3.15.0
- 3.14.1
- 3.13.0
- 3.12.0
- 3.11.4
- 3.4.0
- 3.3.6
- 3.2.0
- 3.1.0
- 3.0.1
- 2.34.4
- 2.33.0
- 2.32.0
- 2.31.0
- 2.30.1
- 2.29.0
- 2.28.1
- 2.27.1
- 2.26.0
- 2.25.2
- 2.24.1
- 2.23.3
- 2.22.1
- 2.21.0
- 2.20.0
- 2.19.0
- 2.18.0
- 2.17.0
- 2.16.1
- 2.15.0
- 2.14.0
- 2.13.1
- 2.12.0
- 2.11.0
- 2.10.0
- 2.9.0
- 2.8.0
- 2.7.0
- 2.6.2
- 2.5.0
- 2.4.0
- 2.3.1
- 2.2.0
- 2.1.0
- 2.0.0
- 1.28.2
- 1.27.2
- 1.26.1
- 1.25.0
- 1.24.0
- 1.23.1
- 1.22.0
- 1.21.0
- 1.20.0
- 1.19.0
- 1.18.0
- 1.17.0
- 1.16.0
3.0.0 Migration Guide
New Required Dependencies
Some of the previously optional dependencies are now required in 3.x
versions of the
library, namely
google-cloud-bigquery-storage
(minimum version 2.0.0
) and pyarrow (minimum
version 3.0.0
).
The behavior of some of the package “extras” has thus also changed:
The
pandas
extra now requires the db-types package.The
bqstorage
extra has been preserved for comaptibility reasons, but it is now a no-op and should be omitted when installing the BigQuery client library.
Before:
$ pip install google-cloud-bigquery[bqstorage]
After:
$ pip install google-cloud-bigquery
- The
bignumeric_type
extra has been removed, asBIGNUMERIC
type is now automatically supported. That extra should thus not be used.
Before:
$ pip install google-cloud-bigquery[bignumeric_type]
After:
$ pip install google-cloud-bigquery
Type Annotations
The library is now type-annotated and declares itself as such. If you use a static
type checker such as mypy
, you might start getting errors in places where
google-cloud-bigquery
package is used.
It is recommended to update your code and/or type annotations to fix these errors, but
if this is not feasible in the short term, you can temporarily ignore type annotations
in google-cloud-bigquery
, for example by using a special # type: ignore
comment:
from google.cloud import bigquery # type: ignore
But again, this is only recommended as a possible short-term workaround if immediately fixing the type check errors in your project is not feasible.
Re-organized Types
The auto-generated parts of the library has been removed, and proto-based types formerly
found in google.cloud.bigquery_v2
have been replaced by the new implementation (but
see the section below).
For example, the standard SQL data types should new be imported from a new location:
Before:
from google.cloud.bigquery_v2 import StandardSqlDataType
from google.cloud.bigquery_v2.types import StandardSqlField
from google.cloud.bigquery_v2.types.standard_sql import StandardSqlStructType
After:
from google.cloud.bigquery import StandardSqlDataType
from google.cloud.bigquery.standard_sql import StandardSqlField
from google.cloud.bigquery.standard_sql import StandardSqlStructType
The TypeKind
enum defining all possible SQL types for schema fields has been renamed
and is not nested anymore under StandardSqlDataType
:
Before:
from google.cloud.bigquery_v2 import StandardSqlDataType
if field_type == StandardSqlDataType.TypeKind.STRING:
...
After:
from google.cloud.bigquery import StandardSqlTypeNames
if field_type == StandardSqlTypeNames.STRING:
...
Issuing queries with Client.create_job
preserves destination table
The Client.create_job
method no longer removes the destination table from a
query job’s configuration. Destination table for the query can thus be
explicitly defined by the user.
Changes to data types when reading a pandas DataFrame
The default dtypes returned by the to_dataframe
method have changed.
Now, the BigQuery
BOOLEAN
data type maps to the pandasboolean
dtype. Previously, this mapped to the pandasbool
dtype when the column did not containNULL
values and the pandasobject
dtype whenNULL
values are present.Now, the BigQuery
INT64
data type maps to the pandasInt64
dtype. Previously, this mapped to the pandasint64
dtype when the column did not containNULL
values and the pandasfloat64
dtype whenNULL
values are present.Now, the BigQuery
DATE
data type maps to the pandasdbdate
dtype, which is provided by the db-dtypes package. If any date value is outside of the range of pandas.Timestamp.min (1677-09-22) and pandas.Timestamp.max (2262-04-11), the data type maps to the pandasobject
dtype. Thedate_as_object
parameter has been removed.Now, the BigQuery
TIME
data type maps to the pandasdbtime
dtype, which is provided by the db-dtypes package.
Changes to data types loading a pandas DataFrame
In the absence of schema information, pandas columns with naive
datetime64[ns]
values, i.e. without timezone information, are recognized and
loaded using the DATETIME
type. On the other hand, for columns with
timezone-aware datetime64[ns, UTC]
values, the TIMESTAMP
type is continued
to be used.
Changes to Model
, Client.get_model
, Client.update_model
, and Client.list_models
The types of several Model
properties have been changed.
Model.feature_columns
now returns a sequence ofgoogle.cloud.bigquery.standard_sql.StandardSqlField
.Model.label_columns
now returns a sequence ofgoogle.cloud.bigquery.standard_sql.StandardSqlField
.Model.model_type
now returns a string.Model.training_runs
now returns a sequence of dictionaries, as recieved from the BigQuery REST API.
Legacy Protocol Buffers Types
For compatibility reasons, the legacy proto-based types still exists as static code and can be imported:
from google.cloud.bigquery_v2 import Model # a sublcass of proto.Message
Mind, however, that importing them will issue a warning, because aside from
being importable, these types are not maintained anymore. They may differ
both from the types in google.cloud.bigquery
, and from the types supported on
the backend.
Maintaining compatibility with google-cloud-bigquery
version 2.0
If you maintain a library or system that needs to support both
google-cloud-bigquery
version 2.x and 3.x, it is recommended that you detect
when version 2.x is in use and convert properties that use the legacy protocol
buffer types, such as Model.training_runs
, into the types used in 3.x.
Call the to_dict
method
on the protocol buffers objects to get a JSON-compatible dictionary.
from google.cloud.bigquery_v2 import Model
training_run: Model.TrainingRun = ...
training_run_dict = training_run.to_dict()
2.0.0 Migration Guide
The 2.0 release of the google-cloud-bigquery
client drops support for Python
versions below 3.6. The client surface itself has not changed, but the 1.x series
will not be receiving any more feature updates or bug fixes. You are thus
encouraged to upgrade to the 2.x series.
If you experience issues or have questions, please file an issue.
Supported Python Versions
WARNING: Breaking change
The 2.0.0 release requires Python 3.6+.
Supported BigQuery Storage Clients
The 2.0.0 release requires BigQuery Storage >= 2.0.0
, which dropped support
for v1beta1
and v1beta2
versions of the BigQuery Storage API. If you want to
use a BigQuery Storage client, it must be the one supporting the v1
API version.
Changed GAPIC Enums Path
WARNING: Breaking change
Generated GAPIC enum types have been moved under types
. Import paths need to be
adjusted.
Before:
from google.cloud.bigquery_v2.gapic import enums
distance_type = enums.Model.DistanceType.COSINE
After:
from google.cloud.bigquery_v2 import types
distance_type = types.Model.DistanceType.COSINE