As of January 1, 2020 this library no longer supports Python 2 on the latest released version.
Library versions released prior to that date will continue to be available. For more information please
visit Python 2 support on Google Cloud.
Source code for google.cloud.firestore_v1.base_client
# Copyright 2017 Google LLC All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Client for interacting with the Google Cloud Firestore API.
This is the base from which all interactions with the API occur.
In the hierarchy of API concepts
* a :class:`~google.cloud.firestore_v1.client.Client` owns a
:class:`~google.cloud.firestore_v1.collection.CollectionReference`
* a :class:`~google.cloud.firestore_v1.client.Client` owns a
:class:`~google.cloud.firestore_v1.document.DocumentReference`
"""
import os
from typing import (
Any,
AsyncGenerator,
Generator,
Iterable,
List,
Optional,
Tuple,
Union,
)
import google.api_core.client_options
import google.api_core.path_template
import grpc # type: ignore
from google.api_core import retry as retries
from google.api_core.gapic_v1 import client_info
from google.auth.credentials import AnonymousCredentials
from google.cloud.client import ClientWithProject # type: ignore
from google.cloud.firestore_v1 import __version__, _helpers, types
from google.cloud.firestore_v1.base_batch import BaseWriteBatch
# Types needed only for Type Hints
from google.cloud.firestore_v1.base_collection import BaseCollectionReference
from google.cloud.firestore_v1.base_document import (
BaseDocumentReference,
DocumentSnapshot,
)
from google.cloud.firestore_v1.base_query import BaseQuery
from google.cloud.firestore_v1.base_transaction import BaseTransaction
from google.cloud.firestore_v1.bulk_writer import BulkWriter, BulkWriterOptions
from google.cloud.firestore_v1.field_path import render_field_path
DEFAULT_DATABASE = "(default)"
"""str: The default database used in a :class:`~google.cloud.firestore_v1.client.Client`."""
_DEFAULT_EMULATOR_PROJECT = "google-cloud-firestore-emulator"
_BAD_OPTION_ERR = (
"Exactly one of ``last_update_time`` or ``exists`` " "must be provided."
)
_BAD_DOC_TEMPLATE: str = (
"Document {!r} appeared in response but was not present among references"
)
_ACTIVE_TXN: str = "There is already an active transaction."
_INACTIVE_TXN: str = "There is no active transaction."
_CLIENT_INFO: Any = client_info.ClientInfo(client_library_version=__version__)
_FIRESTORE_EMULATOR_HOST: str = "FIRESTORE_EMULATOR_HOST"
[docs]class BaseClient(ClientWithProject):
"""Client for interacting with Google Cloud Firestore API.
.. note::
Since the Cloud Firestore API requires the gRPC transport, no
``_http`` argument is accepted by this class.
Args:
project (Optional[str]): The project which the client acts on behalf
of. If not passed, falls back to the default inferred
from the environment.
credentials (Optional[~google.auth.credentials.Credentials]): The
OAuth2 Credentials to use for this client. If not passed, falls
back to the default inferred from the environment.
database (Optional[str]): The database name that the client targets.
For now, :attr:`DEFAULT_DATABASE` (the default value) is the
only valid database.
client_info (Optional[google.api_core.gapic_v1.client_info.ClientInfo]):
The client info used to send a user-agent string along with API
requests. If ``None``, then default info will be used. Generally,
you only need to set this if you're developing your own library
or partner tool.
client_options (Union[dict, google.api_core.client_options.ClientOptions]):
Client options used to set user options on the client. API Endpoint
should be set through client_options.
"""
SCOPE = (
"https://www.googleapis.com/auth/cloud-platform",
"https://www.googleapis.com/auth/datastore",
)
"""The scopes required for authenticating with the Firestore service."""
_firestore_api_internal = None
_database_string_internal = None
_rpc_metadata_internal = None
def __init__(
self,
project=None,
credentials=None,
database=None,
client_info=_CLIENT_INFO,
client_options=None,
) -> None:
database = database or DEFAULT_DATABASE
# NOTE: This API has no use for the _http argument, but sending it
# will have no impact since the _http() @property only lazily
# creates a working HTTP object.
self._emulator_host = os.getenv(_FIRESTORE_EMULATOR_HOST)
if self._emulator_host is not None:
if credentials is None:
credentials = AnonymousCredentials()
if project is None:
# extract project from env var, or use system default
project = (
os.getenv("GOOGLE_CLOUD_PROJECT")
or os.getenv("GCLOUD_PROJECT")
or _DEFAULT_EMULATOR_PROJECT
)
super(BaseClient, self).__init__(
project=project,
credentials=credentials,
client_options=client_options,
_http=None,
)
self._client_info = client_info
if client_options:
if isinstance(client_options, dict):
client_options = google.api_core.client_options.from_dict(
client_options
)
self._client_options = client_options
self._database = database
def _firestore_api_helper(self, transport, client_class, client_module) -> Any:
"""Lazy-loading getter GAPIC Firestore API.
Returns:
The GAPIC client with the credentials of the current client.
"""
if self._firestore_api_internal is None:
# Use a custom channel.
# We need this in order to set appropriate keepalive options.
if self._emulator_host is not None:
channel = self._emulator_channel(transport)
else:
channel = transport.create_channel(
self._target,
credentials=self._credentials,
options={"grpc.keepalive_time_ms": 30000}.items(),
)
self._transport = transport(host=self._target, channel=channel)
self._firestore_api_internal = client_class(
transport=self._transport, client_options=self._client_options
)
client_module._client_info = self._client_info
return self._firestore_api_internal
def _emulator_channel(self, transport):
"""
Creates an insecure channel to communicate with the local emulator.
If credentials are provided the token is extracted and added to the
headers. This supports local testing of firestore rules if the credentials
have been created from a signed custom token.
:return: grpc.Channel or grpc.aio.Channel
"""
# Insecure channels are used for the emulator as secure channels
# cannot be used to communicate on some environments.
# https://github.com/googleapis/python-firestore/issues/359
# Default the token to a non-empty string, in this case "owner".
token = "owner"
if (
self._credentials is not None
and getattr(self._credentials, "id_token", None) is not None
):
token = self._credentials.id_token
options = [("Authorization", f"Bearer {token}")]
if "GrpcAsyncIOTransport" in str(transport.__name__):
return grpc.aio.insecure_channel(self._emulator_host, options=options)
else:
return grpc.insecure_channel(self._emulator_host, options=options)
def _target_helper(self, client_class) -> str:
"""Return the target (where the API is).
Eg. "firestore.googleapis.com"
Returns:
str: The location of the API.
"""
if self._emulator_host is not None:
return self._emulator_host
elif self._client_options and self._client_options.api_endpoint:
return self._client_options.api_endpoint
else:
return client_class.DEFAULT_ENDPOINT
@property
def _database_string(self):
"""The database string corresponding to this client's project.
This value is lazy-loaded and cached.
Will be of the form
``projects/{project_id}/databases/{database_id}``
but ``database_id == '(default)'`` for the time being.
Returns:
str: The fully-qualified database string for the current
project. (The default database is also in this string.)
"""
if self._database_string_internal is None:
db_str = google.api_core.path_template.expand(
"projects/{project}/databases/{database}",
project=self.project,
database=self._database,
)
self._database_string_internal = db_str
return self._database_string_internal
@property
def _rpc_metadata(self):
"""The RPC metadata for this client's associated database.
Returns:
Sequence[Tuple(str, str)]: RPC metadata with resource prefix
for the database associated with this client.
"""
if self._rpc_metadata_internal is None:
self._rpc_metadata_internal = _helpers.metadata_with_prefix(
self._database_string
)
if self._emulator_host is not None:
# The emulator requires additional metadata to be set.
self._rpc_metadata_internal.append(("authorization", "Bearer owner"))
return self._rpc_metadata_internal
def collection(self, *collection_path) -> BaseCollectionReference[BaseQuery]:
raise NotImplementedError
def collection_group(self, collection_id: str) -> BaseQuery:
raise NotImplementedError
def _get_collection_reference(
self, collection_id: str
) -> BaseCollectionReference[BaseQuery]:
"""Checks validity of collection_id and then uses subclasses collection implementation.
Args:
collection_id (str) Identifies the collections to query over.
Every collection or subcollection with this ID as the last segment of its
path will be included. Cannot contain a slash.
Returns:
The created collection.
"""
if "/" in collection_id:
raise ValueError(
"Invalid collection_id "
+ collection_id
+ ". Collection IDs must not contain '/'."
)
return self.collection(collection_id)
def document(self, *document_path) -> BaseDocumentReference:
raise NotImplementedError
[docs] def bulk_writer(self, options: Optional[BulkWriterOptions] = None) -> BulkWriter:
"""Get a BulkWriter instance from this client.
Args:
:class:`@google.cloud.firestore_v1.bulk_writer.BulkWriterOptions`:
Optional control parameters for the
:class:`@google.cloud.firestore_v1.bulk_writer.BulkWriter` returned.
Returns:
:class:`@google.cloud.firestore_v1.bulk_writer.BulkWriter`:
A utility to efficiently create and save many `WriteBatch` instances
to the server.
"""
return BulkWriter(client=self, options=options)
def _document_path_helper(self, *document_path) -> List[str]:
"""Standardize the format of path to tuple of path segments and strip the database string from path if present.
Args:
document_path (Tuple[str, ...]): Can either be
* A single ``/``-delimited path to a document
* A tuple of document path segments
"""
path = _path_helper(document_path)
base_path = self._database_string + "/documents/"
joined_path = _helpers.DOCUMENT_PATH_DELIMITER.join(path)
if joined_path.startswith(base_path):
joined_path = joined_path[len(base_path) :]
return joined_path.split(_helpers.DOCUMENT_PATH_DELIMITER)
def recursive_delete(
self,
reference: Union[BaseCollectionReference[BaseQuery], BaseDocumentReference],
bulk_writer: Optional["BulkWriter"] = None, # type: ignore
) -> int:
raise NotImplementedError
[docs] @staticmethod
def field_path(*field_names: str) -> str:
"""Create a **field path** from a list of nested field names.
A **field path** is a ``.``-delimited concatenation of the field
names. It is used to represent a nested field. For example,
in the data
.. code-block:: python
data = {
'aa': {
'bb': {
'cc': 10,
},
},
}
the field path ``'aa.bb.cc'`` represents the data stored in
``data['aa']['bb']['cc']``.
Args:
field_names: The list of field names.
Returns:
str: The ``.``-delimited field path.
"""
return render_field_path(field_names)
[docs] @staticmethod
def write_option(
**kwargs,
) -> Union[_helpers.ExistsOption, _helpers.LastUpdateOption]:
"""Create a write option for write operations.
Write operations include :meth:`~google.cloud.DocumentReference.set`,
:meth:`~google.cloud.DocumentReference.update` and
:meth:`~google.cloud.DocumentReference.delete`.
One of the following keyword arguments must be provided:
* ``last_update_time`` (:class:`google.protobuf.timestamp_pb2.\
Timestamp`): A timestamp. When set, the target document must
exist and have been last updated at that time. Protobuf
``update_time`` timestamps are typically returned from methods
that perform write operations as part of a "write result"
protobuf or directly.
* ``exists`` (:class:`bool`): Indicates if the document being modified
should already exist.
Providing no argument would make the option have no effect (so
it is not allowed). Providing multiple would be an apparent
contradiction, since ``last_update_time`` assumes that the
document **was** updated (it can't have been updated if it
doesn't exist) and ``exists`` indicate that it is unknown if the
document exists or not.
Args:
kwargs (Dict[str, Any]): The keyword arguments described above.
Raises:
TypeError: If anything other than exactly one argument is
provided by the caller.
Returns:
:class:`~google.cloud.firestore_v1.client.WriteOption`:
The option to be used to configure a write message.
"""
if len(kwargs) != 1:
raise TypeError(_BAD_OPTION_ERR)
name, value = kwargs.popitem()
if name == "last_update_time":
return _helpers.LastUpdateOption(value)
elif name == "exists":
return _helpers.ExistsOption(value)
else:
extra = "{!r} was provided".format(name)
raise TypeError(_BAD_OPTION_ERR, extra)
def _prep_get_all(
self,
references: list,
field_paths: Iterable[str] = None,
transaction: BaseTransaction = None,
retry: retries.Retry = None,
timeout: float = None,
) -> Tuple[dict, dict, dict]:
"""Shared setup for async/sync :meth:`get_all`."""
document_paths, reference_map = _reference_info(references)
mask = _get_doc_mask(field_paths)
request = {
"database": self._database_string,
"documents": document_paths,
"mask": mask,
"transaction": _helpers.get_transaction_id(transaction),
}
kwargs = _helpers.make_retry_timeout_kwargs(retry, timeout)
return request, reference_map, kwargs
def get_all(
self,
references: list,
field_paths: Iterable[str] = None,
transaction: BaseTransaction = None,
retry: retries.Retry = None,
timeout: float = None,
) -> Union[
AsyncGenerator[DocumentSnapshot, Any], Generator[DocumentSnapshot, Any, Any]
]:
raise NotImplementedError
def _prep_collections(
self,
retry: retries.Retry = None,
timeout: float = None,
) -> Tuple[dict, dict]:
"""Shared setup for async/sync :meth:`collections`."""
request = {"parent": "{}/documents".format(self._database_string)}
kwargs = _helpers.make_retry_timeout_kwargs(retry, timeout)
return request, kwargs
def collections(
self,
retry: retries.Retry = None,
timeout: float = None,
) -> Union[
AsyncGenerator[BaseCollectionReference[BaseQuery], Any],
Generator[BaseCollectionReference[BaseQuery], Any, Any],
]:
raise NotImplementedError
def batch(self) -> BaseWriteBatch:
raise NotImplementedError
def transaction(self, **kwargs) -> BaseTransaction:
raise NotImplementedError
def _reference_info(references: list) -> Tuple[list, dict]:
"""Get information about document references.
Helper for :meth:`~google.cloud.firestore_v1.client.Client.get_all`.
Args:
references (List[.DocumentReference, ...]): Iterable of document
references.
Returns:
Tuple[List[str, ...], Dict[str, .DocumentReference]]: A two-tuple of
* fully-qualified documents paths for each reference in ``references``
* a mapping from the paths to the original reference. (If multiple
``references`` contains multiple references to the same document,
that key will be overwritten in the result.)
"""
document_paths = []
reference_map = {}
for reference in references:
doc_path = reference._document_path
document_paths.append(doc_path)
reference_map[doc_path] = reference
return document_paths, reference_map
def _get_reference(document_path: str, reference_map: dict) -> BaseDocumentReference:
"""Get a document reference from a dictionary.
This just wraps a simple dictionary look-up with a helpful error that is
specific to :meth:`~google.cloud.firestore.client.Client.get_all`, the
**public** caller of this function.
Args:
document_path (str): A fully-qualified document path.
reference_map (Dict[str, .DocumentReference]): A mapping (produced
by :func:`_reference_info`) of fully-qualified document paths to
document references.
Returns:
.DocumentReference: The matching reference.
Raises:
ValueError: If ``document_path`` has not been encountered.
"""
try:
return reference_map[document_path]
except KeyError:
msg = _BAD_DOC_TEMPLATE.format(document_path)
raise ValueError(msg)
def _parse_batch_get(
get_doc_response: types.BatchGetDocumentsResponse,
reference_map: dict,
client: BaseClient,
) -> DocumentSnapshot:
"""Parse a `BatchGetDocumentsResponse` protobuf.
Args:
get_doc_response (~google.cloud.proto.firestore.v1.\
firestore.BatchGetDocumentsResponse): A single response (from
a stream) containing the "get" response for a document.
reference_map (Dict[str, .DocumentReference]): A mapping (produced
by :func:`_reference_info`) of fully-qualified document paths to
document references.
client (:class:`~google.cloud.firestore_v1.client.Client`):
A client that has a document factory.
Returns:
[.DocumentSnapshot]: The retrieved snapshot.
Raises:
ValueError: If the response has a ``result`` field (a oneof) other
than ``found`` or ``missing``.
"""
result_type = get_doc_response._pb.WhichOneof("result")
if result_type == "found":
reference = _get_reference(get_doc_response.found.name, reference_map)
data = _helpers.decode_dict(get_doc_response.found.fields, client)
snapshot = DocumentSnapshot(
reference,
data,
exists=True,
read_time=get_doc_response.read_time,
create_time=get_doc_response.found.create_time,
update_time=get_doc_response.found.update_time,
)
elif result_type == "missing":
reference = _get_reference(get_doc_response.missing, reference_map)
snapshot = DocumentSnapshot(
reference,
None,
exists=False,
read_time=get_doc_response.read_time,
create_time=None,
update_time=None,
)
else:
raise ValueError(
"`BatchGetDocumentsResponse.result` (a oneof) had a field other "
"than `found` or `missing` set, or was unset"
)
return snapshot
def _get_doc_mask(field_paths: Iterable[str]) -> Optional[types.common.DocumentMask]:
"""Get a document mask if field paths are provided.
Args:
field_paths (Optional[Iterable[str, ...]]): An iterable of field
paths (``.``-delimited list of field names) to use as a
projection of document fields in the returned results.
Returns:
Optional[google.cloud.firestore_v1.types.common.DocumentMask]: A mask
to project documents to a restricted set of field paths.
"""
if field_paths is None:
return None
else:
return types.DocumentMask(field_paths=field_paths)
def _path_helper(path: tuple) -> Tuple[str]:
"""Standardize path into a tuple of path segments.
Args:
path (Tuple[str, ...]): Can either be
* A single ``/``-delimited path
* A tuple of path segments
"""
if len(path) == 1:
return path[0].split(_helpers.DOCUMENT_PATH_DELIMITER)
else:
return path