Skip to content

Latest commit

 

History

History
243 lines (174 loc) · 20.3 KB

README.md

File metadata and controls

243 lines (174 loc) · 20.3 KB

Target-BigQuery

Actions Status License: MIT Code style: black

A rare 💎 you have stumbled upon

target-bigquery is a Singer target for BigQuery.

It is the most versatile target for BigQuery. Extremely performant, resource efficient, and fast in all configurations enabling 20 different ingestion patterns. Denormalized variants indicate data is unpacked during load with a resultant schema in BigQuery based on the tap schema. Non-denormalized means we have a fixed schema which loads all data into an unstructured JSON column. They are both useful patterns. The latter allowing BigQuery to work with schemaless or rapidly changing sources such as MongoDB instantly, while the former is more performant and convenient to start modeling quickly.

Patterns 🛠 (more details below)

  • Batch Job, Denormalized, Overwrite
  • Batch Job, Denormalized, Upsert
  • Batch Job, Denormalized, Append
  • Batch Job, Fixed Schema, Overwrite
  • Batch Job, Fixed Schema, Append
  • GCS Staging Data Lake, Denormalized, Overwrite
  • GCS Staging Data Lake, Denormalized, Upsert
  • GCS Staging Data Lake, Denormalized, Append
  • GCS Staging Data Lake, Fixed Schema, Overwrite
  • GCS Staging Data Lake, Fixed Schema, Append
  • Storage Write API, Denormalized, Overwrite
  • Storage Write API, Denormalized, Upsert
  • Storage Write API, Denormalized, Append
  • Storage Write API, Fixed Schema, Overwrite
  • Storage Write API, Fixed Schema, Append
  • Legacy Streaming API, Denormalized, Overwrite
  • Legacy Streaming API, Denormalized, Upsert
  • Legacy Streaming API, Denormalized, Append
  • Legacy Streaming API, Fixed Schema, Overwrite
  • Legacy Streaming API, Fixed Schema, Append

Installation 📈

The package on pypi is named z3-target-bigquery but the executable it ships with is simply target-bigquery. This allows me to release work without concerns of naming conflicts on the package index.

# Use pipx or pip
pipx install z3-target-bigquery
# Verify it is installed
target-bigquery --version

Usage Notes

Denormalization

So denormalized is a loaded term, so lets clarify here what we mean by it:

In the context of JSON objects, denormalization refers to the process of flattening a hierarchical or nested JSON object into a simpler, more "denormalized" structure that can be easier to work with for certain use cases.

Denormalized=False (default) means we load all data into a single JSON column. This means all access requires an accessor such as SELECT data.my_column FROM table instead of SELECT my_column FROM table. Hence the term denormalized is relative to the data column which is the default for this target. This is a tradeoff between convenience/resilience and performance. This is the default because most resilient. IE a load will never fail due to a schema change or an invalid schema. You can always denormalize later via dbt. However, it is not the most performant and slower to transform. If your tap has a high quality and consistent schema, denormalization is the way to go to get the best performance and start modeling quickly.

Denormalized=True means we unpack the data into a schema which is derived from the tap schema. It does not mean we will flatten the data. There is a separate option for flattening. We will convert arrays to repeated fields and records to structs. All top level keys will end up as columns which is was you might expect from more typical targets.

Resolver Versions

There are 2 resolver versions. The config option schema_resolver_version lets you select which version you want to use. This versioning exists because we want to support evolving how we resolve schemas whilst not creating breaking changes for long-time users dependent on how a schema is resolved. The default is 1 which behaves very similarly to existing flavors of target-bigquery. It works well enough but has plenty of edge cases where it simply cannot resolve valid jsonschemas to a bq schema. The new version 2 is much more robust and will resolve most, if not all schemas due to it falling back to JSON when in doubt. You must opt-in to this version by setting schema_resolver_version: 2 in your config.

Overwrite vs Append

Sometimes you want to overwrite a table on every load. This can be achieved by either setting overwrite: true which will full refresh ALL tables or setting overwrite: [table1, table2, table_*_other, !table_v1_other] which will only overwrite the specified tables and supports pattern matching. This is useful if you have a table which is a lookup table and you want to overwrite it on every load. You can also set overwrite: false which will append to the table. This is the default behavior.

Upsert (aka Merge)

If you want to merge data into a table, you can set merge: true which will use the MERGE statement to upsert data. This supports pattern matching like the above setting. It requires denormalized: true takes precedence over overwrite. It will only work on tables which have a primary key as defined by the key_properties sent by the tap. There is a supporting config option called dedupe_before_upsert which will dedupe the data before upserting. This is useful if you are replicating data which has a primary key but is not unique. This occurs when you are replicating data from a source which has a primary key but does not enforce it. This is the case for MongoDB. It can also happen when moving data from a data lake in S3/GCS to a database. This is not the default behavior because it is slower and requires more resources.

Features ✨

  • Autoscaling self-healing worker pool using either threads (default) or multiprocessing, configurable by the user for the fastest possible data ingestion. Particularly when leveraging colocated compute in GCP.
  • Denormalized load pattern where data is unpacked in flight into a statically typed BigQuery schema derived from the input stream json schemas.
  • Fix schema load pattern where all data is loaded into a JSON column which has been GA in BigQuery since mid 2022.
  • Autogenerated VIEW support for fixed schema load patterns which essentially overlays a statically typed schema allowing you to get the best of both worlds when using fixed schema ingestion.
  • JIT compilation of protobuf schemas allowing the Storage Write API to use a denormalized load pattern.
  • BATCH message support 😎

Load Patterns 🏎

  • Batch Load Job ingestion pattern using in memory compression (fixed schema + denormalized)
  • Storage Write API ingestion pattern using gRPC and protocol buffers supporting both streaming and batch patterns. Capable of JIT compilation of BQ schemas to protobuf to enable denormalized loads of input structures only known at runtime. (fixed schema + denormalized 🎉)
  • GCS Staging ingestion pattern using in memory compression and a GCS staging layer which generates a well organized data lake which backs the data warehouse providing additional failsafes and data sources (fixed schema + denormalized)
  • Legacy Streaming ingestion pattern which emphasizes simplicity and fast start up / tear down. I would highly recommend the storage write API instead unless data volume is small (fixed schema + denormalized)

Choosing between denormalized and fixed schema (JSON support)? 🙇🏾‍♂️

The gap between the methods is closed due in part to the target's ability to automatically generating a VIEW which will unpack (or rather provide typing as a more accurate take) a JSON based ingestion source for you. Unless operating at tens of millions of rows with JSON objects containing multiple hundreds of keys, its quite performant. Particularly if accessing a small subset of keys. It does however fall off (quite hard) given enough scale as I mentioned (I've pushed it to the limits). Denormalized is recommended for high volume where the schema is fairly consistent. Fixed is recommended for lower volume, inconsistent schemas or for taps which are inherently schemaless in which case its the ideal (only...) logical pattern. Fixed schema can also be used for taps which routinely break down on BQ due to json schemas being inexpressible in a static way (ie patternProperties, additionalProperties...)

Old Header (still true, here for posterity) 🧪

This is the first truly unstructured sink for BigQuery leveraging the recent GA feature in BigQuery for JSON support. This allows this target to load from essentially any tap regardless of the quality or explicitness of its jsonschema. Observations in existing taps note things such as patternProperties used in jsonschema objects which break down on all existing BigQuery taps due to the previous need for strong typing. Also taps such as MongoDB which inherently deal with unstructured data are seamlessly enabled by this target without klutzy collection scraping of a sample of records which we hope are repesentative of all documents.

Built with the Meltano Target SDK.

Configuration 🔨

Settings

First a valid example to give context to the below including a nested key example (denoted via a . in the setting path) as seen with column_name_transforms.snake_case

{
    "project": "my-bq-project",
    "method": "storage_write_api",
    "denormalized": true,
    "credentials_path": "...",
    "dataset": "my_dataset",
    "location": "us-central1",
    "batch_size": 500,
    "column_name_transforms": {
      "snake_case": true
    }
}
Setting Required Default Description
credentials_path False None The path to a gcp credentials json file.
credentials_json False None A JSON string of your service account JSON file.
project True None The target GCP project to materialize data into.
dataset True None The target dataset to materialize data into.
location False US The target dataset location to materialize data into. Applies also to the GCS bucket if using gcs_stage load method.
batch_size False 500 The maximum number of rows to send in a single batch to the worker. This should be configured based on load method. For storage_write_api and streaming_insert it should be <=500, for the LoadJob sinks, it can be much higher, ie >100,000
timeout False 600 Default timeout for batch_job and gcs_stage derived LoadJobs.
fail_fast False True Fail the entire load job if any row fails to insert.
denormalized False False Determines whether to denormalize the data before writing to BigQuery. A false value will write data using a fixed JSON column based schema, while a true value will write data using a dynamic schema derived from the tap.
method True storage_write_api The method to use for writing to BigQuery. Must be one of batch_job, storage_write_api, gcs_stage, streaming_insert
generate_view False False Determines whether to generate a view based on the SCHEMA message parsed from the tap. Only valid if denormalized=false meaning you are using the fixed JSON column based schema.
upsert False False Determines if we should upsert. Defaults to false. A value of true will write to a temporary table and then merge into the target table (upsert). This requires the target table to be unique on the key properties. A value of false will write to the target table directly (append). A value of an array of strings will evaluate the strings in order using fnmatch. At the end of the array, the value of the last match will be used. If not matched, the default value is false (append).
overwrite False False Determines if the target table should be overwritten on load. Defaults to false. A value of true will write to a temporary table and then overwrite the target table inside a transaction (so it is safe). A value of false will write to the target table directly (append). A value of an array of strings will evaluate the strings in order using fnmatch. At the end of the array, the value of the last match will be used. If not matched, the default value is false. This is mutually exclusive with the upsert option. If both are set, upsert will take precedence.
dedupe_before_upsert False False This option is only used if upsert is enabled for a stream. The selection criteria for the stream's candidacy is the same as upsert. If the stream is marked for deduping before upsert, we will create a _session scoped temporary table during the merge transaction to dedupe the ingested records. This is useful for streams that are not unique on the key properties during an ingest but are unique in the source system. Data lake ingestion is often a good example of this where the same unique record may exist in the lake at different points in time from different extracts.
bucket False None The GCS bucket to use for staging data. Only used if method is gcs_stage.
cluster_on_key_properties False 0 Determines whether to cluster on the key properties from the tap. Defaults to false. When false, clustering will be based on _sdc_batched_at instead.
partition_granularity False "month" Indicates the granularity of the created table partitioning scheme which is based on _sdc_batched_at. By default the granularity is monthly. Must be one of: "hour", "day", "month", "year".
column_name_transforms.lower False None Lowercase column names.
column_name_transforms.quote False None Quote column names in any generated DDL.
column_name_transforms.add_underscore_when_invalid False None Add an underscore to the column name if it starts with a digit to make it valid.
column_name_transforms.snake_case False None Snake case all incoming column names. Does not apply to fixed schema loads but does apply to the view auto-generated over them.
options.storage_write_batch_mode False None By default, we use the default stream (Committed mode) in the storage_write_api load method which results in streaming records which are immediately available and is generally fastest. If this is set to true, we will use the application created streams (pending mode) to transactionally batch data on STATE messages and at end of pipe.
options.process_pool False None By default we use an autoscaling threadpool to write to BigQuery. If set to true, we will use a process pool.
options.max_workers False None By default, each sink type has a preconfigured max worker pool limit. This sets an override for maximum number of workers in the pool.
schema_resolver_version False 1 The version of the schema resolver to use. Defaults to 1. Version 2 uses JSON as a fallback during denormalization. This only has an effect if denormalized=true
stream_maps False None Config object for stream maps capability. For more information check out Stream Maps.
stream_map_config False None User-defined config values to be used within map expressions.
flattening_enabled False None 'True' to enable schema flattening and automatically expand nested properties.
flattening_max_depth False None The max depth to flatten schemas.

A full list of supported settings and capabilities is available by running: target-bigquery --about

Configure using environment variables ✏️

This Singer target will automatically import any environment variables within the working directory's .env if the --config=ENV is provided, such that config values will be considered if a matching environment variable is set either in the terminal context or in the .env file.

Source Authentication and Authorization 👮🏽‍♂️

Authenticate via service account key file or Application Default Credentials (ADC) https://cloud.google.com/bigquery/docs/authentication

Capabilities ✨

  • about
  • stream-maps
  • schema-flattening
  • batch

Usage 👷‍♀️

You can easily run target-bigquery by itself or in a pipeline using Meltano.

Executing the Target Directly 🚧

target-bigquery --version
target-bigquery --help
# Test using the "Carbon Intensity" sample:
tap-carbon-intensity | target-bigquery --config /path/to/target-bigquery-config.json

Developer Resources 👩🏼‍💻

Initialize your Development Environment

pipx install poetry
poetry install

Create and Run Tests

Create tests within the target_bigquery/tests subfolder and then run:

poetry run pytest

You can also test the target-bigquery CLI interface directly using poetry run:

poetry run target-bigquery --help

Testing with Meltano

Note: This target will work in any Singer environment and does not require Meltano. Examples here are for convenience and to streamline end-to-end orchestration scenarios.

Next, install Meltano (if you haven't already) and any needed plugins:

# Install meltano
pipx install meltano
# Initialize meltano within this directory
cd target-bigquery
meltano install

Now you can test and orchestrate using Meltano:

# Test invocation:
meltano invoke target-bigquery --version
# OR run a test `elt` pipeline with the Carbon Intensity sample tap:
meltano elt tap-carbon-intensity target-bigquery

SDK Dev Guide

See the dev guide for more instructions on how to use the Meltano SDK to develop your own Singer taps and targets.