100100 HDFS_KERB_TICKET ,
101101 HDFS_PORT ,
102102 HDFS_USER ,
103- PYARROW_USE_LARGE_TYPES_ON_READ ,
104103 S3_ACCESS_KEY_ID ,
105104 S3_ANONYMOUS ,
106105 S3_CONNECT_TIMEOUT ,
179178from pyiceberg .utils .config import Config
180179from pyiceberg .utils .datetime import millis_to_datetime
181180from pyiceberg .utils .decimal import unscaled_to_decimal
182- from pyiceberg .utils .deprecated import deprecation_message
183181from pyiceberg .utils .properties import get_first_property_value , property_as_bool , property_as_int
184182from pyiceberg .utils .singleton import Singleton
185183from pyiceberg .utils .truncate import truncate_upper_bound_binary_string , truncate_upper_bound_text_string
@@ -1756,14 +1754,6 @@ def to_table(self, tasks: Iterable[FileScanTask]) -> pa.Table:
17561754 (pa .Table .from_batches ([batch ]) for batch in itertools .chain ([first_batch ], batches )), promote_options = "permissive"
17571755 )
17581756
1759- if property_as_bool (self ._io .properties , PYARROW_USE_LARGE_TYPES_ON_READ , False ):
1760- deprecation_message (
1761- deprecated_in = "0.10.0" ,
1762- removed_in = "0.11.0" ,
1763- help_message = f"Property `{ PYARROW_USE_LARGE_TYPES_ON_READ } ` will be removed." ,
1764- )
1765- result = result .cast (arrow_schema )
1766-
17671757 return result
17681758
17691759 def to_record_batches (self , tasks : Iterable [FileScanTask ]) -> Iterator [pa .RecordBatch ]:
@@ -1872,7 +1862,6 @@ class ArrowProjectionVisitor(SchemaWithPartnerVisitor[pa.Array, pa.Array | None]
18721862 _file_schema : Schema
18731863 _include_field_ids : bool
18741864 _downcast_ns_timestamp_to_us : bool
1875- _use_large_types : bool | None
18761865 _projected_missing_fields : dict [int , Any ]
18771866 _allow_timestamp_tz_mismatch : bool
18781867
@@ -1881,26 +1870,17 @@ def __init__(
18811870 file_schema : Schema ,
18821871 downcast_ns_timestamp_to_us : bool = False ,
18831872 include_field_ids : bool = False ,
1884- use_large_types : bool | None = None ,
18851873 projected_missing_fields : dict [int , Any ] = EMPTY_DICT ,
18861874 allow_timestamp_tz_mismatch : bool = False ,
18871875 ) -> None :
18881876 self ._file_schema = file_schema
18891877 self ._include_field_ids = include_field_ids
18901878 self ._downcast_ns_timestamp_to_us = downcast_ns_timestamp_to_us
1891- self ._use_large_types = use_large_types
18921879 self ._projected_missing_fields = projected_missing_fields
18931880 # When True, allows projecting timestamptz (UTC) to timestamp (no tz).
18941881 # Allowed for reading (aligns with Spark); disallowed for writing to enforce Iceberg spec's strict typing.
18951882 self ._allow_timestamp_tz_mismatch = allow_timestamp_tz_mismatch
18961883
1897- if use_large_types is not None :
1898- deprecation_message (
1899- deprecated_in = "0.10.0" ,
1900- removed_in = "0.11.0" ,
1901- help_message = "Argument `use_large_types` will be removed from ArrowProjectionVisitor" ,
1902- )
1903-
19041884 def _cast_if_needed (self , field : NestedField , values : pa .Array ) -> pa .Array :
19051885 file_field = self ._file_schema .find_field (field .field_id )
19061886
@@ -1949,8 +1929,6 @@ def _cast_if_needed(self, field: NestedField, values: pa.Array) -> pa.Array:
19491929 target_schema = schema_to_pyarrow (
19501930 promote (file_field .field_type , field .field_type ), include_field_ids = self ._include_field_ids
19511931 )
1952- if self ._use_large_types is False :
1953- target_schema = _pyarrow_schema_ensure_small_types (target_schema )
19541932 return values .cast (target_schema )
19551933
19561934 return values
0 commit comments