| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import logging |
| from contextlib import contextmanager |
| from google.api_core.exceptions import GoogleAPICallError |
|
|
| logger = logging.getLogger(__name__) |
| try: |
| from opentelemetry import trace |
| from opentelemetry.instrumentation.utils import http_status_to_status_code |
| from opentelemetry.trace.status import Status |
|
|
| HAS_OPENTELEMETRY = True |
| _warned_telemetry = True |
|
|
| except ImportError: |
| HAS_OPENTELEMETRY = False |
| _warned_telemetry = False |
|
|
| _default_attributes = { |
| "db.system": "BigQuery" |
| } |
|
|
|
|
| @contextmanager |
| def create_span(name, attributes=None, client=None, job_ref=None): |
| """Creates a ContextManager for a Span to be exported to the configured exporter. |
| If no configuration exists yields None. |
| |
| Args: |
| name (str): Name that will be set for the span being created |
| attributes (Optional[dict]): |
| Additional attributes that pertain to |
| the specific API call (i.e. not a default attribute) |
| client (Optional[google.cloud.bigquery.client.Client]): |
| Pass in a Client object to extract any attributes that may be |
| relevant to it and add them to the created spans. |
| job_ref (Optional[google.cloud.bigquery.job._AsyncJob]) |
| Pass in a _AsyncJob object to extract any attributes that may be |
| relevant to it and add them to the created spans. |
| |
| Yields: |
| opentelemetry.trace.Span: Yields the newly created Span. |
| |
| Raises: |
| google.api_core.exceptions.GoogleAPICallError: |
| Raised if a span could not be yielded or issue with call to |
| OpenTelemetry. |
| """ |
| global _warned_telemetry |
| final_attributes = _get_final_span_attributes(attributes, client, job_ref) |
| if not HAS_OPENTELEMETRY: |
| if not _warned_telemetry: |
| logger.debug( |
| "This service is instrumented using OpenTelemetry. " |
| "OpenTelemetry or one of its components could not be imported; " |
| "please add compatible versions of opentelemetry-api and " |
| "opentelemetry-instrumentation packages in order to get BigQuery " |
| "Tracing data." |
| ) |
| _warned_telemetry = True |
|
|
| yield None |
| return |
| tracer = trace.get_tracer(__name__) |
|
|
| |
| with tracer.start_as_current_span(name=name, attributes=final_attributes) as span: |
| try: |
| yield span |
| except GoogleAPICallError as error: |
| if error.code is not None: |
| span.set_status(Status(http_status_to_status_code(error.code))) |
| raise |
|
|
|
|
| def _get_final_span_attributes(attributes=None, client=None, job_ref=None): |
| """Compiles attributes from: client, job_ref, user-provided attributes. |
| |
| Attributes from all of these sources are merged together. Note the |
| attributes are added sequentially based on perceived order of precedence: |
| i.e. attributes added last may overwrite attributes added earlier. |
| |
| Args: |
| attributes (Optional[dict]): |
| Additional attributes that pertain to |
| the specific API call (i.e. not a default attribute) |
| |
| client (Optional[google.cloud.bigquery.client.Client]): |
| Pass in a Client object to extract any attributes that may be |
| relevant to it and add them to the final_attributes |
| |
| job_ref (Optional[google.cloud.bigquery.job._AsyncJob]) |
| Pass in a _AsyncJob object to extract any attributes that may be |
| relevant to it and add them to the final_attributes. |
| |
| Returns: dict |
| """ |
|
|
| collected_attributes = _default_attributes.copy() |
|
|
| if client: |
| collected_attributes.update(_set_client_attributes(client)) |
| if job_ref: |
| collected_attributes.update(_set_job_attributes(job_ref)) |
| if attributes: |
| collected_attributes.update(attributes) |
|
|
| final_attributes = {k: v for k, v in collected_attributes.items() if v is not None} |
| return final_attributes |
|
|
|
|
| def _set_client_attributes(client): |
| return {"db.name": client.project, "location": client.location} |
|
|
|
|
| def _set_job_attributes(job_ref): |
| job_attributes = { |
| "db.name": job_ref.project, |
| "job_id": job_ref.job_id, |
| "state": job_ref.state, |
| } |
|
|
| job_attributes["hasErrors"] = job_ref.error_result is not None |
|
|
| if job_ref.created is not None: |
| job_attributes["timeCreated"] = job_ref.created.isoformat() |
|
|
| if job_ref.started is not None: |
| job_attributes["timeStarted"] = job_ref.started.isoformat() |
|
|
| if job_ref.ended is not None: |
| job_attributes["timeEnded"] = job_ref.ended.isoformat() |
|
|
| if job_ref.location is not None: |
| job_attributes["location"] = job_ref.location |
|
|
| if job_ref.parent_job_id is not None: |
| job_attributes["parent_job_id"] = job_ref.parent_job_id |
|
|
| if job_ref.num_child_jobs is not None: |
| job_attributes["num_child_jobs"] = job_ref.num_child_jobs |
|
|
| total_bytes_billed = getattr(job_ref, "total_bytes_billed", None) |
| if total_bytes_billed is not None: |
| job_attributes["total_bytes_billed"] = total_bytes_billed |
|
|
| total_bytes_processed = getattr(job_ref, "total_bytes_processed", None) |
| if total_bytes_processed is not None: |
| job_attributes["total_bytes_processed"] = total_bytes_processed |
|
|
| return job_attributes |
|
|