Quantcast
Channel: Active questions tagged python - Stack Overflow
Viewing all articles
Browse latest Browse all 23131

dask's read_parquet reads in files that should be filtered out when loading schema

$
0
0

I'm running into an issue where I'm saving different parquet files with slightly different schema, but which have shared partition columns. I have created the following as a minimal reproducible example:

from dask import dataframe as ddimport pandas as pdimport shutildef save_parquet():    df1 = pd.DataFrame({"A": [1], "B": [1]})    df2 = pd.DataFrame({"A": [2], "C": [2]})    df1.to_parquet("test.parquet", partition_cols=["A"])    df2.to_parquet("test.parquet", partition_cols=["A"])def load_parquet():    filters = [[        ("A", "==", 2)    ]]    ddf = dd.read_parquet("test.parquet", columns=["A", "C"], filters=filters)def main():    save_parquet()    load_parquet()if __name__=="__main__":    main()

Running the above causes the following exception:

Traceback (most recent call last):  File "/home/.../.local/lib/python3.8/site-packages/dask/backends.py", line 133, in wrapper    return func(*args, **kwargs)  File "/home/.../.local/lib/python3.8/site-packages/dask/dataframe/io/parquet/core.py", line 578, in read_parquet    meta, index, columns = set_index_columns(meta, index, columns, auto_index_allowed)  File "/home/.../.local/lib/python3.8/site-packages/dask/dataframe/io/parquet/core.py", line 1487, in set_index_columns    raise ValueError(ValueError: The following columns were not found in the dataset {'C'}The following columns were found Index(['B', 'A'], dtype='object')The above exception was the direct cause of the following exception:Traceback (most recent call last):  File "mre.py", line 26, in <module>    main()  File "mre.py", line 23, in main    load_parquet()  File "mre.py", line 15, in load_parquet    ddf = dd.read_parquet("test.parquet", columns=["A", "C"], filters=filters)  File "/home/.../.local/lib/python3.8/site-packages/dask/backends.py", line 135, in wrapper    raise type(e)(ValueError: An error occurred while calling the read_parquet method registered to the pandas backend.Original Message: The following columns were not found in the dataset {'C'}The following columns were found Index(['B', 'A'], dtype='object')

My expectation would be that the ("A", "==", 2) filter should stop us from loading the schema from df1, and regardless of whether it loads df1 or not, that it should be able to find the "C" column from df2. Am I missing something here?

Changing the columns field to columns=["A", "B"] successfully reads in the data, so it feels what I'm trying to do should be possible somehow.

This post indicates that read_parquet reads the schema from the first parquet file it encounters, but that you can specify a schema to avoid this.

Specifying a schema like

import pyarrow as pa...ddf = dd.read_parquet("test.parquet", columns=["A", "C"], filters=filters, schema=pa.schema({"A": pa.int64(), "C": pa.int64()}))

still triggers the exception.

Specifying a schema without specifying columns doesn't trigger the exception, but returns a dask dataframe without the "C" column (seemingly regardless of what was in the schema):

>>> print(ddf.columns)Index(['B', 'A'], dtype='object')

Is there a way to prevent read_parquet from using the .parquet files that should be filtered out?


Viewing all articles
Browse latest Browse all 23131

Trending Articles



<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>