@@ -31,7 +31,7 @@ use crate::arrow::array::RecordBatch;
3131use crate :: arrow:: datatypes:: { Fields , Schema , SchemaRef } ;
3232use crate :: datasource:: file_format:: file_compression_type:: FileCompressionType ;
3333use crate :: datasource:: file_format:: write:: get_writer_schema;
34- use crate :: datasource:: physical_plan:: parquet:: { can_expr_be_pushed_down_with_schemas, get_reader_options_config_or_default, MetadataFetcher , ReaderOptionsConfig } ;
34+ use crate :: datasource:: physical_plan:: parquet:: { can_expr_be_pushed_down_with_schemas, get_reader_options_config_or_default, get_reader_options_customizer , MetadataFetcher } ;
3535use crate :: datasource:: physical_plan:: parquet:: source:: ParquetSource ;
3636use crate :: datasource:: physical_plan:: { FileSink , FileSinkConfig } ;
3737use crate :: datasource:: statistics:: { create_max_min_accs, get_col_stats} ;
@@ -432,7 +432,7 @@ impl FileFormat for ParquetFormat {
432432
433433 async fn create_physical_plan (
434434 & self ,
435- _state : & dyn Session ,
435+ state : & dyn Session ,
436436 conf : FileScanConfig ,
437437 filters : Option < & Arc < dyn PhysicalExpr > > ,
438438 ) -> Result < Arc < dyn ExecutionPlan > > {
@@ -451,7 +451,7 @@ impl FileFormat for ParquetFormat {
451451 metadata_size_hint = Some ( metadata) ;
452452 }
453453
454- let mut source = ParquetSource :: new ( self . options . clone ( ) , ReaderOptionsConfig :: noop ( ) ) ; // TODO upgrade DF
454+ let mut source = ParquetSource :: new ( self . options . clone ( ) , get_reader_options_customizer ( state . config ( ) ) ) ;
455455
456456 if let Some ( predicate) = predicate {
457457 source = source. with_predicate ( Arc :: clone ( & conf. file_schema ) , predicate) ;
0 commit comments