What happens?
TypeError: write_parquet(): incompatible function arguments. The following argument types are supported:
1. (self: _duckdb.DuckDBPyRelation, file_name: str, *, compression: object = None, field_ids: object = None, row_group_size_bytes: object = None, row_group_size: object = None, overwrite: object = None, per_thread_output: object = None, use_tmp_file: object = None, partition_by: object = None, write_partition_columns: object = None, append: object = None, filename_pattern: object = None, file_size_bytes: object = None) -> None
while invoking the write parquet function with the params that are accepted and documented in https://duckdb.org/docs/current/sql/statements/copy#parquet-options
To Reproduce
import duckdb
con = duckdb.connect()
con.sql("select 1 as test").write_parquet("test.parquet", row_groups_per_file=2)
OS:
Ubuntu Linux x86_64
DuckDB Package Version:
1.5.0
Python Version:
1.13.1
Full Name:
Niccolò Cantù
Affiliation:
Norwegian Institute for Nature Research (NINA)
What is the latest build you tested with? If possible, we recommend testing with the latest nightly build.
I have tested with a stable release
Did you include all relevant data sets for reproducing the issue?
Not applicable - the reproduction does not require a data set
Did you include all code required to reproduce the issue?
Did you include all relevant configuration to reproduce the issue?
What happens?
while invoking the write parquet function with the params that are accepted and documented in https://duckdb.org/docs/current/sql/statements/copy#parquet-options
To Reproduce
OS:
Ubuntu Linux x86_64
DuckDB Package Version:
1.5.0
Python Version:
1.13.1
Full Name:
Niccolò Cantù
Affiliation:
Norwegian Institute for Nature Research (NINA)
What is the latest build you tested with? If possible, we recommend testing with the latest nightly build.
I have tested with a stable release
Did you include all relevant data sets for reproducing the issue?
Not applicable - the reproduction does not require a data set
Did you include all code required to reproduce the issue?
Did you include all relevant configuration to reproduce the issue?