Table of Contents
This section discusses different design aspects of kartothek Cubes Functionality. Not that many decisions here are heavily bound to the Kartothek format and the features it provides. Also, Kartothek Cube must support the usage patterns we have seen and provide a clear API (in technical and semantical terms). It especially is no allrounder solution and we do NOT want to support all possible use cases.
When mapping multiple parts (tables or datasets) to Kartothek, we have the following options:
Multi-Table: As used by Klee1, use a single dataset w/ multiple tables.
Multi-Dataset: Use multiple datasets each w/ a single table.
We use multiple Kartothek datasets instead of multiple tables (like Klee1 did) for the following reasons:
Kartothek discourages unbalanced tables, e.g. when Kartothek Cube writes a seed table and an enrich step is only done for a part of the data (e.g. the present and future). This worked for Klee1 because we have used a way simpler format (based on Kartothek V2) and ignored the recommendation of the Kartothek main author.
OSS tools like Dask.DataFrame do not support any form of multi-table datasets and the current Kartothek format is just a workaround to make multi-table datasets look like multiple regular datasets.
The Kartothek`developers want to deprecate multi-table `Kartothek datasets in the future.
Kartothek needs a central metadata file for performance reason (listing blob containers w/ massive amounts of blobs is very slow) which is shared between tables. So running 2 tasks that create 2 completely unrelated tables would still result in write conflicts. Using separate datasets solves this issue.
Using multiple datasets allows users to copy, backup and delete them separately.
Index structures are bound to datasets which feels more consistent than the former solution where you did not know which table was meant by an index entry.
There are two ways how a cube could be created out of Kartothek datasets:
UUID Namespace: specify a UUID prefix and the base cube on existing of datasets (if a dataset w/ a given prefix exists, it is part of the cube)
Cube Metadata File: write another metadata file, similar to what Kartothek is doing but like a “dataset of datasets”
By using another metadata file, we would introduce a potential write conflict, one that was solved by using multiple datasets instead of multiple tables. Also, it would make copying, deleting and backup up datasets in an independent fashion more difficult.
Using an UUID prefix an implicit existing does not has these issues, but requires us to list the store content for dataset discovery. While listing all store keys can be really slow, listing w/ a path separator is really quick, especially if the number of “folders” and “top-level files” is small, which is given using Kartothek V4 datasets. Also, it allows us to load external datasets (like ones produces by other systems) as cubes and therefore simplifies interoperability.
Since all datasets that match a given prefix are part of the cube, they all must be valid in terms of the Format.
There are multiple ways to store data to the blob store:
Single, Large Parquet Files: Just one giant Parquet file for everything.
Explicit Physical Partitions: Split the data into partitions described by dedicated columns.
Automatic Partitioning: Leave the load-balancing entirely to kartothek_cube.
We have decided for explicit physical partitions since we have seen that this data model works well for our current data flow. It allows quick and efficient re-partitioning to allow row-based, group-based, and timeseries-based data processing steps, while keeping the technical complexity rather low (compared to an automatic + dynamic partitioning). It also maps well to multiple backends we plan to use.
Using a single, large Parquet file would scale during the read path due to predicate pushdown and using well-thought chunk sizes, but either require a write path that supports writing individual chunks (currently not available and unlikely to be supported anytime soon) or one fat-node at the end of every write pipeline which is inefficient and even impossible w/ our current architecture.
In a future version of Klee, there may be a way to get the data automatically partitioned, maybe even w/ some feedback-loop based approach (submit tasks, watch memory/CPU consumption, adjust, re-submit).
Note that while physical partitions are explicit, their semantic impact should be small. They are an optimization and may speed up load and store operations, but cubes w/ different partition columns but built out of the same data should behave the same (except some query features like re-partioning may be differ due to missing indices).
We are aware of two ways to design the update granularity:
Partition-wise: Entire partitions can overwrite old physical partitions. Deletion operations are partition-wise.
Row-wise: The entire cube behaves like one large, virtual DataFrame and the user handles rows. Physical partitions are just an optimization.
While the row-wise approach has the nice property that the physical partitioning is just an optimization, it is complex to implement and a major performance problem, especially when many transaction were written to the cube. This is due to the fact that the query planner cannot inspect the cube cells from each parquet file without reading it and therefore either needs a very expensive planning phase with loads of IO operations or it cannot prune data early, leading to massive IO in the execution phase. So we decided for partition-wise IO, which turned out to be quite simple to implement.
There are multiple options regarding the mapping of dataset columns to DataFrame columns:
No Mapping: do not change column names but prohibit collisions of payload columns (i.e. columns that are neither dimension nor partition columns)
Namespace all Columns: e.g. dimension and partition columns have now prefix, but payload columns have the form
<dataset ID>__<column name>
Namespace on Demand: only prefix in case of a name collision, similar to what DataFrame.merge is doing.
Since Kartothek Cube is intended to use in production, “Namespace on Demand” is not an option since it may result in hard to debug runtime errors. “Namespace all Columns” is a very stable option, but would require every part of our data processing pipelines to know which dataset produces which column. Currently, this is transparent and columns can be moved from one stage to another w/o resulting to larger code changes. We would like to keep this nice behavior, so we went for “No Mapping”.
When data is stored in multiple parts (tables or datasets), the question is how to expose it to the user during read operations:
Seperate DataFrames: Conditions are group-by operations are applied to the individual parts and no join operations are performed by kartothek_cube.
Seed-Based Join: Mark a single part as seed which provides the groundtruth regarding cells (i.e. unique dimension entries) in the cube, all other parts are just additional columns.
Fully Configurable Join Order: Leave it to the user to configure the join order (this was done in an early version of Klee1).
Separate DataFrames would give the user full control, but would also force them to create load of boilerplate code, likely resulting in another framework on top of kartothek_cube. This would contradict any KISS approach we try to take here. Also it makes reasoning about conditions and partition-by parameters more difficult since it is not always clear how these effects cross-influence different parts of the cube.
Using a fully configurable was tried in Klee1, but it turned out that many users do not want to care about the consequences of this complex beast. Also, it makes predicate pushdown and efficient index operations more difficult to implement, especially since the core of Kartothek Cube is based on Pandas which lacks proper NULL-handling.
Finally, we have decided for a seed-based system some time ago in Klee1 and our users are happy and know what to expect. It is also easy to teach, good to implement and test, and it matches the semantic of our data processing pipelines (groundtruth from an initial external source, subsequent enrichments w/ additional columns on top of it.)
There are two variants of the seed-based system:
Enforced: Cells in non-seed datasets must be present in the seed dataset. This is enforced during write operations.
Lazy: The seed semantic is only enforced during queries.
We have decided for the lazy approach, since it better supports independent copies and backups of datasets and also simplifies some of our processing pipelines (e.g. geolocation data can blindly be fetched for too many locaations and dates.)
This section describes how Kartothek must be structured to be consumed by kartothek_cube.
An abstract cube is described by the following attributes:
UUID Prefix: A common prefix for UUIDs of all datasets that are part of the cube.
Dimension Columns: Which orthogonal dimensions form the cube. Hence, every cell described by these columns is unique.
Partition Columns: Columns that describe how data is partitioned when written to the blob store. These columns will form the Kartothek
Seed Dataset: Which dataset forms the ground truth regarding the set of cells in the cube.
Index Columns: Which non-dimension and non-partition columns should also be indexed.
All Kartothek datasets that are part of a cube must fulfill the following criteria:
Kartothek Dataset UUID:
'<UUID Prefix>++<Dataset UUID>'. E.g. for a cube called
'my_cube'and a dataset called
'weather', the UUID that is used in Kartothek is
Metadata Version: 4
Kartothek Tables: Only a single one called
Other Datasets: Can be anything.
Partition Labels: The user has no ability set the partition labels, instead the default Kartothek UUID4 generation mechanism is used.
The following index structures must be present (additional indices will be ignored):
Partition Indices: According to the partition keys described above.
Explicit Secondary Indices: For all index columns. For the seed dataset also for all dimension columns. Additional indices may exist and can be used by the query engine.
Kartothek Cube allows the user to specify per-dataset metadata. Furthermore, the following entries are added by default to every dataset:
On top of what Kartothek is doing, the following properties of preserved DataFrames will be ensured:
all column names are unicode strings (
str); that especially excludes integer-based column names
DataFrame indices are range indices starting at 0 with a step size of 1; this is equivalent to DataFrame.reset_index
values are sorted by dimension columns (if present) in the order given by cube specification.
Examples in docstrings, README and specs should use real-world column names (like
Examples in pytest should use simplified column names:
CLI examples are produced using
kartothek_cube --color=always ... | terminal-to-html > out.html with terminal-to-html and
are wrapped into the following code snipped:
.. raw:: html <pre> ... </pre>