Skip to content

Commit 4c979c8

Browse files
MINOR: [Docs][Python] Mention explicit None partitioning o[ptions for pyarrow.parquet.read_table
1 parent 48d5151 commit 4c979c8

File tree

2 files changed

+6
-5
lines changed

2 files changed

+6
-5
lines changed

python/pyarrow/dataset.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -542,7 +542,7 @@ def parquet_dataset(metadata_path, schema=None, filesystem=None, format=None,
542542
format : ParquetFileFormat
543543
An instance of a ParquetFileFormat if special options needs to be
544544
passed.
545-
partitioning : Partitioning, PartitioningFactory, str, list of str
545+
partitioning : Partitioning, PartitioningFactory, str, list of str, optional
546546
The partitioning scheme specified with the ``partitioning()``
547547
function. A flavor string can be used as shortcut, and with a list of
548548
field names a DirectoryPartitioning will be inferred.
@@ -638,7 +638,7 @@ def dataset(source, schema=None, format=None, filesystem=None,
638638
examples below.
639639
Note that the URIs on Windows must follow 'file:///C:...' or
640640
'file:/C:...' patterns.
641-
partitioning : Partitioning, PartitioningFactory, str, list of str
641+
partitioning : Partitioning, PartitioningFactory, str, list of str, optional
642642
The partitioning scheme specified with the ``partitioning()``
643643
function. A flavor string can be used as shortcut, and with a list of
644644
field names a DirectoryPartitioning will be inferred.

python/pyarrow/parquet/core.py

+4-3
Original file line numberDiff line numberDiff line change
@@ -1163,13 +1163,14 @@ def _get_pandas_index_columns(keyvalues):
11631163
buffer_size : int, default 0
11641164
If positive, perform read buffering when deserializing individual
11651165
column chunks. Otherwise IO calls are unbuffered.
1166-
partitioning : pyarrow.dataset.Partitioning or str or list of str, \
1166+
partitioning : pyarrow.dataset.Partitioning or str or list of str or None, \
11671167
default "hive"
11681168
The partitioning scheme for a partitioned dataset. The default of "hive"
11691169
assumes directory names with key=value pairs like "/year=2009/month=11".
11701170
In addition, a scheme like "/2009/11" is also supported, in which case
1171-
you need to specify the field names or a full schema. See the
1172-
``pyarrow.dataset.partitioning()`` function for more details."""
1171+
you need to specify the field names or a full schema. If no partitioning
1172+
is used, pass ``None``.
1173+
See the ``pyarrow.dataset.partitioning()`` function for more details."""
11731174

11741175

11751176
_parquet_dataset_example = """\

0 commit comments

Comments
 (0)