GithubHelp home page GithubHelp logo

Comments (1)

trymzet avatar trymzet commented on August 14, 2024

Ah it turns out advise users to run SHOW SESSION. For reference, here is the list of values I got:

Name Default Type Description
adaptive_partial_aggregation_enabled TRUE boolean When enabled, partial aggregation might be adaptively turned off when it does not provide any performance gain
adaptive_partial_aggregation_min_rows 100000 bigint Minimum number of processed rows before partial aggregation might be adaptively turned off
adaptive_partial_aggregation_unique_rows_ratio_threshold 0.8 double Ratio between aggregation output and input rows above which partial aggregation might be adaptively turned off
aggregation_operator_unspill_memory_limit 4MB varchar How much memory should be allocated per aggregation operator in unspilling process
collect_plan_statistics_for_all_queries FALSE boolean Collect plan statistics for non-EXPLAIN queries
colocated_join FALSE boolean Experimental: Use a colocated join when possible
concurrent_lifespans_per_task 0 integer Experimental: Run a fixed number of groups concurrently for eligible JOINs
default_filter_factor_enabled FALSE boolean use a default filter factor for unknown filters in a filter node
dictionary_aggregation FALSE boolean Enable optimization for aggregations on dictionaries
distributed_index_join FALSE boolean Distribute index joins on join keys instead of executing inline
distributed_sort TRUE boolean Parallelize sort across multiple nodes
dynamic_schedule_for_grouped_execution FALSE boolean Experimental: Use dynamic schedule for grouped execution when possible
enable_coordinator_dynamic_filters_distribution TRUE boolean Enable distribution of dynamic filters from coordinator to all workers
enable_dynamic_filtering TRUE boolean Enable dynamic filtering
enable_intermediate_aggregations FALSE boolean Enable the use of intermediate aggregations
enable_large_dynamic_filters FALSE boolean Enable collection of large dynamic filters
enable_stats_calculator TRUE boolean Enable statistics calculator
exchange_compression FALSE boolean Enable compression in exchanges
execution_policy phased varchar Policy used for scheduling query tasks
fault_tolerant_execution_max_task_split_count 256 integer Maximal number of splits for a single fault tolerant task (count based)
fault_tolerant_execution_min_task_split_count 16 integer Minimal number of splits for a single fault tolerant task (count based)
fault_tolerant_execution_target_task_input_size 1GB varchar Target size in bytes of all task inputs for a single fault tolerant task
fault_tolerant_execution_target_task_split_count 16 integer Target number of splits for a single fault tolerant task (split weight aware)
fault_tolerant_execution_task_memory 4GB varchar Estimated amount of memory a single task will use when task level retries are used; value is used allocating nodes for tasks execution
fault_tolerant_execution_task_memory_estimation_quantile 0.9 double What quantile of memory usage of completed tasks to look at when estimating memory usage for upcoming tasks
fault_tolerant_execution_task_memory_growth_factor 3 double Factor by which estimated task memory is increased if task execution runs out of memory; value is used allocating nodes for tasks execution
filter_and_project_min_output_page_row_count 256 integer Experimental: Minimum output page row count for filter and project operators
filter_and_project_min_output_page_size 500kB varchar Experimental: Minimum output page size for filter and project operators
filter_conjunction_independence_factor 0.75 double Scales the strength of independence assumption for selectivity estimates of the conjunction of multiple filters
grouped_execution FALSE boolean Use grouped execution when possible
hash_partition_count 100 integer Number of partitions for distributed joins and aggregations
hide_inaccessible_columns FALSE boolean When enabled non-accessible columns are silently filtered from results from SELECT * statements
ignore_downstream_preferences FALSE boolean Ignore Parent's PreferredProperties in AddExchange optimizer
ignore_stats_calculator_failures TRUE boolean Ignore statistics calculator failures
incremental_hash_array_load_factor_enabled TRUE boolean Use smaller load factor for small hash arrays in order to improve performance
initial_splits_per_node 2 integer The number of splits each node will run per task, initially
iterative_optimizer_timeout 3.00m varchar Timeout for plan optimization in iterative optimizer
join_distribution_type AUTOMATIC varchar Join distribution type. Possible values: [BROADCAST, PARTITIONED, AUTOMATIC]
join_max_broadcast_table_size 100MB varchar Maximum estimated size of a table that can be broadcast when using automatic join type selection
join_multi_clause_independence_factor 0.25 double Scales the strength of independence assumption for selectivity estimates of multi-clause joins
join_reordering_strategy AUTOMATIC varchar Join reordering strategy. Possible values: [NONE, ELIMINATE_CROSS_JOINS, AUTOMATIC]
late_materialization FALSE boolean Experimental: Use late materialization (including WorkProcessor pipelines)
max_drivers_per_task   integer Maximum number of drivers per task
max_partial_top_n_memory 16MB varchar Max memory size for partial Top N aggregations. This can be turned off by setting it with '0'.
max_recursion_depth 10 integer Maximum recursion depth for recursive common table expression
max_reordered_joins 9 integer The maximum number of joins to reorder as one group in cost-based join reordering
max_tasks_waiting_for_node_per_stage 5 integer Maximum possible number of tasks waiting for node allocation per stage before scheduling of new tasks for stage is paused
max_unacknowledged_splits_per_task 500 integer Maximum number of leaf splits awaiting delivery to a given task
merge_project_with_values TRUE boolean Inline project expressions into values
non_estimatable_predicate_approximation_enabled TRUE boolean Approximate the cost of filters which cannot be accurately estimated even with complete statistics
omit_datetime_type_precision FALSE boolean Omit precision when rendering datetime type names with default precision
optimize_duplicate_insensitive_joins TRUE boolean Optimize duplicate insensitive joins
optimize_hash_generation TRUE boolean Compute hash codes for distribution, joins, and aggregations early in query plan
optimize_metadata_queries FALSE boolean Enable optimization for metadata queries
optimize_mixed_distinct_aggregations FALSE boolean Optimize mixed non-distinct and distinct aggregations
optimize_top_n_ranking TRUE boolean Use top N ranking optimization
parse_decimal_literals_as_double FALSE boolean Parse decimal literals as DOUBLE instead of DECIMAL
predicate_pushdown_use_table_properties TRUE boolean Use table properties in predicate pushdown
prefer_partial_aggregation TRUE boolean Prefer splitting aggregations into partial and final stages
prefer_streaming_operators FALSE boolean Prefer source table layouts that produce streaming operators
preferred_write_partitioning_min_number_of_partitions 50 integer Use preferred write partitioning when the number of written partitions exceeds the configured threshold
push_aggregation_through_outer_join TRUE boolean Allow pushing aggregations below joins
push_partial_aggregation_through_join FALSE boolean Push partial aggregations below joins
push_table_write_through_union TRUE boolean Parallelize writes when using UNION ALL in queries that write data
query_max_cpu_time 1000000000.00d varchar Maximum CPU time of a query
query_max_execution_time 100.00d varchar Maximum execution time of a query
query_max_planning_time 10.00m varchar Maximum planning time of a query
query_max_run_time 100.00d varchar Maximum run time of a query (includes the queueing time)
query_max_scan_physical_bytes   varchar Maximum scan physical bytes of a query
query_priority 1 integer The priority of queries. Larger numbers are higher priority
query_retry_attempts 4 integer Maximum number of query retry attempts
redistribute_writes TRUE boolean Force parallel distributed writes
required_workers_count 1 integer Minimum number of active workers that must be available before the query will start
required_workers_max_wait_time 5.00m varchar Maximum time to wait for minimum number of workers before the query is failed
resource_overcommit FALSE boolean Use resources which are not guaranteed to be available to the query
retry_delay_scale_factor 2 double Maximum delay before initiating a retry attempt. Delay increases exponentially for each subsequent attempt starting from 'retry_initial_delay'
retry_initial_delay 10.00s varchar Initial delay before initiating a retry attempt. Delay increases exponentially for each subsequent attempt up to 'retry_max_delay'
retry_max_delay 1.00m varchar Maximum delay before initiating a retry attempt. Delay increases exponentially for each subsequent attempt starting from 'retry_initial_delay'
rewrite_filtering_semi_join_to_inner_join TRUE boolean Rewrite semi join in filtering context to inner join
scale_writers TRUE boolean Scale out writers based on throughput (use minimum necessary)
skip_redundant_sort TRUE boolean Skip redundant sort operations
spatial_join TRUE boolean Use spatial index for spatial join when possible
spatial_partitioning_table_name   varchar Name of the table containing spatial partitioning scheme
spill_enabled FALSE boolean Enable spilling
split_concurrency_adjustment_interval 100.00ms varchar Experimental: Interval between changes to the number of concurrent splits per node
statistics_cpu_timer_enabled TRUE boolean Experimental: Enable cpu time tracking for automatic column statistics collection on write
statistics_precalculation_for_pushdown_enabled TRUE boolean Enable statistics precalculation for pushdown
table_scan_node_partitioning_min_bucket_to_task_ratio 0.5 double Min table scan bucket to task ratio for which plan will be adopted to node pre-partitioned tables
task_concurrency 1 integer Default number of local parallel jobs per worker
task_retry_attempts_overall 2147483647 integer Maximum number of task retry attempts overall
task_retry_attempts_per_task 4 integer Maximum number of task retry attempts per single task
task_share_index_loading FALSE boolean Share index join lookups and caching within a task
task_writer_count 1 integer Default number of local parallel table writer jobs per worker
use_legacy_window_filter_pushdown FALSE boolean Use legacy window filter pushdown optimizer
use_mark_distinct TRUE boolean Implement DISTINCT aggregations using MarkDistinct
use_preferred_write_partitioning TRUE boolean Use preferred write partitioning
use_table_scan_node_partitioning TRUE boolean Adapt plan to node pre-partitioned tables
writer_min_size 32MB varchar Target minimum size of writer output when scaling writers
hive.bucket_execution_enabled TRUE boolean Enable bucket-aware execution: only use a single worker per bucket
hive.collect_column_statistics_on_write TRUE boolean Enables automatic column level statistics collection on write
hive.compression_codec GZIP varchar Compression codec to use when writing files. Possible values: [NONE, SNAPPY, LZ4, ZSTD, GZIP]
hive.create_empty_bucket_files FALSE boolean Create empty files for buckets that have no data
hive.dynamic_filtering_wait_timeout 0.00m varchar Duration to wait for completion of dynamic filters during split generation
hive.experimental_parquet_optimized_writer_enabled FALSE boolean Experimental: Enable optimized writer
hive.force_local_scheduling FALSE boolean Only schedule splits on workers colocated with data node
hive.hive_storage_format ORC varchar Default storage format for new tables or partitions. Possible values: [ORC, PARQUET, AVRO, RCBINARY, RCTEXT, SEQUENCEFILE, JSON, TEXTFILE, CSV]
hive.hive_views_legacy_translation FALSE boolean Use legacy Hive view translation mechanism
hive.ignore_absent_partitions FALSE boolean Ignore partitions when the file system location does not exist rather than failing the query.
hive.ignore_corrupted_statistics FALSE boolean Experimental: Ignore corrupted statistics rather than failing
hive.insert_existing_partitions_behavior APPEND varchar Behavior on insert existing partitions; this session property doesn't control behavior on insert existing unpartitioned table
hive.minimum_assigned_split_weight 0.05 double Minimum assigned split weight when size based split weighting is enabled
hive.non_transactional_optimize_enabled FALSE boolean Enable OPTIMIZE table procedure
hive.optimize_mismatched_bucket_count FALSE boolean Experimental: Enable optimization to avoid shuffle when bucket count is compatible but not the same
hive.optimize_symlink_listing TRUE boolean Optimize listing for SymlinkTextFormat tables with files in a single directory
hive.orc_bloom_filters_enabled FALSE boolean ORC: Enable bloom filters for predicate pushdown
hive.orc_lazy_read_small_ranges TRUE boolean Experimental: ORC: Read small file segments lazily
hive.orc_max_buffer_size 8MB varchar ORC: Maximum size of a single read
hive.orc_max_merge_distance 1MB varchar ORC: Maximum size of gap between two reads to merge into a single read
hive.orc_max_read_block_size 16MB varchar ORC: Soft max size of Trino blocks produced by ORC reader
hive.orc_nested_lazy_enabled TRUE boolean Experimental: ORC: Lazily read nested data
hive.orc_optimized_writer_max_dictionary_memory 16MB varchar ORC: Max dictionary memory
hive.orc_optimized_writer_max_stripe_rows 10000000 integer ORC: Max stripe row count
hive.orc_optimized_writer_max_stripe_size 64MB varchar ORC: Max stripe size
hive.orc_optimized_writer_min_stripe_size 32MB varchar ORC: Min stripe size
hive.orc_optimized_writer_validate FALSE boolean ORC: Force all validation for files
hive.orc_optimized_writer_validate_mode BOTH varchar ORC: Level of detail in ORC validation. Possible values: [HASHED, DETAILED, BOTH]
hive.orc_optimized_writer_validate_percentage 0 double ORC: sample percentage for validation for files
hive.orc_stream_buffer_size 8MB varchar ORC: Size of buffer for streaming reads
hive.orc_string_statistics_limit 64B varchar ORC: Maximum size of string statistics; drop if exceeding
hive.orc_tiny_stripe_threshold 8MB varchar ORC: Threshold below which an ORC stripe or file will read in its entirety
hive.orc_use_column_names FALSE boolean ORC: Access ORC columns using names from the file
hive.parallel_partitioned_bucketed_writes TRUE boolean Improve parallelism of partitioned and bucketed table writes
hive.parquet_ignore_statistics FALSE boolean Ignore statistics from Parquet to allow querying files with corrupted or incorrect statistics
hive.parquet_max_read_block_size 16MB varchar Parquet: Maximum size of a block to read
hive.parquet_use_column_index TRUE boolean Use Parquet column index
hive.parquet_use_column_names TRUE boolean Parquet: Access Parquet columns using names from the file
hive.parquet_writer_batch_size 10000 integer Parquet: Maximum number of rows passed to the writer in each batch
hive.parquet_writer_block_size 134217728B varchar Parquet: Writer block size
hive.parquet_writer_page_size 1048576B varchar Parquet: Writer page size
hive.partition_statistics_sample_size 100 integer Maximum sample size of the partitions column statistics
hive.projection_pushdown_enabled TRUE boolean Projection push down enabled for hive
hive.propagate_table_scan_sorting_properties FALSE boolean Use sorted table layout to generate more efficient execution plans. May lead to incorrect results if files are not sorted as per table definition.
hive.query_partition_filter_required FALSE boolean Require filter on partition column
hive.query_partition_filter_required_schemas [] array(varchar) List of schemas for which filter on partition column is enforced.
hive.rcfile_optimized_writer_validate FALSE boolean RCFile: Validate writer files
hive.respect_table_format TRUE boolean Write new partitions using table format rather than default storage format
hive.s3_select_pushdown_enabled FALSE boolean S3 Select pushdown enabled
hive.size_based_split_weights_enabled TRUE boolean Enable estimating split weights based on size in bytes
hive.sorted_writing_enabled TRUE boolean Enable writing to bucketed sorted tables
hive.statistics_enabled TRUE boolean Expose table statistics
hive.target_max_file_size 1GB varchar Target maximum size of written files; the actual size may be larger
hive.temporary_staging_directory_enabled TRUE boolean Should use temporary staging directory for write operations
hive.temporary_staging_directory_path /tmp/presto-${USER} varchar Temporary staging directory location
hive.timestamp_precision MILLISECONDS varchar Precision for timestamp columns in Hive tables. Possible values: [MILLISECONDS, MICROSECONDS, NANOSECONDS]
hive.validate_bucketing TRUE boolean Verify that data is bucketed correctly when reading
mysql.aggregation_pushdown_enabled TRUE boolean Enable aggregation pushdown
mysql.decimal_default_scale 0 integer Default decimal scale for mapping unspecified and exceeding precision decimals. Not used when decimal_mapping is set to STRICT
mysql.decimal_mapping STRICT varchar Decimal mapping for unspecified and exceeding precision decimals. STRICT skips them. ALLOW_OVERFLOW requires setting proper decimal scale and rounding mode. Possible values: [STRICT, ALLOW_OVERFLOW]
mysql.decimal_rounding_mode UNNECESSARY varchar Rounding mode for mapping unspecified and exceeding precision decimals. Not used when decimal_mapping is set to STRICT. Possible values: [UP, DOWN, CEILING, FLOOR, HALF_UP, HALF_DOWN, HALF_EVEN, UNNECESSARY]
mysql.domain_compaction_threshold 32 integer Maximum ranges to allow in a tuple domain without simplifying it
mysql.join_pushdown_automatic_max_join_to_tables_ratio 1.25 double If estimated join output size is greater than or equal to ratio * sum of table sizes, then join pushdown will not be performed
mysql.join_pushdown_automatic_max_table_size   varchar Maximum table size to be considered for join pushdown
mysql.join_pushdown_enabled TRUE boolean Enable join pushdown
mysql.join_pushdown_strategy AUTOMATIC varchar Join pushdown strategy. Possible values: [EAGER, AUTOMATIC]
mysql.non_transactional_insert FALSE boolean Do not use temporary table on insert to table
mysql.topn_pushdown_enabled TRUE boolean Enable TopN pushdown
mysql.unsupported_type_handling IGNORE varchar Unsupported type handling strategy. Possible values: [IGNORE, CONVERT_TO_VARCHAR]
mysql.write_batch_size 1000 integer Maximum number of rows to write in a single batch

from dbt-trino.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.