Comments (1)
Ah it turns out advise users to run SHOW SESSION. For reference, here is the list of values I got:
Name | Default | Type | Description |
---|---|---|---|
adaptive_partial_aggregation_enabled | TRUE | boolean | When enabled, partial aggregation might be adaptively turned off when it does not provide any performance gain |
adaptive_partial_aggregation_min_rows | 100000 | bigint | Minimum number of processed rows before partial aggregation might be adaptively turned off |
adaptive_partial_aggregation_unique_rows_ratio_threshold | 0.8 | double | Ratio between aggregation output and input rows above which partial aggregation might be adaptively turned off |
aggregation_operator_unspill_memory_limit | 4MB | varchar | How much memory should be allocated per aggregation operator in unspilling process |
collect_plan_statistics_for_all_queries | FALSE | boolean | Collect plan statistics for non-EXPLAIN queries |
colocated_join | FALSE | boolean | Experimental: Use a colocated join when possible |
concurrent_lifespans_per_task | 0 | integer | Experimental: Run a fixed number of groups concurrently for eligible JOINs |
default_filter_factor_enabled | FALSE | boolean | use a default filter factor for unknown filters in a filter node |
dictionary_aggregation | FALSE | boolean | Enable optimization for aggregations on dictionaries |
distributed_index_join | FALSE | boolean | Distribute index joins on join keys instead of executing inline |
distributed_sort | TRUE | boolean | Parallelize sort across multiple nodes |
dynamic_schedule_for_grouped_execution | FALSE | boolean | Experimental: Use dynamic schedule for grouped execution when possible |
enable_coordinator_dynamic_filters_distribution | TRUE | boolean | Enable distribution of dynamic filters from coordinator to all workers |
enable_dynamic_filtering | TRUE | boolean | Enable dynamic filtering |
enable_intermediate_aggregations | FALSE | boolean | Enable the use of intermediate aggregations |
enable_large_dynamic_filters | FALSE | boolean | Enable collection of large dynamic filters |
enable_stats_calculator | TRUE | boolean | Enable statistics calculator |
exchange_compression | FALSE | boolean | Enable compression in exchanges |
execution_policy | phased | varchar | Policy used for scheduling query tasks |
fault_tolerant_execution_max_task_split_count | 256 | integer | Maximal number of splits for a single fault tolerant task (count based) |
fault_tolerant_execution_min_task_split_count | 16 | integer | Minimal number of splits for a single fault tolerant task (count based) |
fault_tolerant_execution_target_task_input_size | 1GB | varchar | Target size in bytes of all task inputs for a single fault tolerant task |
fault_tolerant_execution_target_task_split_count | 16 | integer | Target number of splits for a single fault tolerant task (split weight aware) |
fault_tolerant_execution_task_memory | 4GB | varchar | Estimated amount of memory a single task will use when task level retries are used; value is used allocating nodes for tasks execution |
fault_tolerant_execution_task_memory_estimation_quantile | 0.9 | double | What quantile of memory usage of completed tasks to look at when estimating memory usage for upcoming tasks |
fault_tolerant_execution_task_memory_growth_factor | 3 | double | Factor by which estimated task memory is increased if task execution runs out of memory; value is used allocating nodes for tasks execution |
filter_and_project_min_output_page_row_count | 256 | integer | Experimental: Minimum output page row count for filter and project operators |
filter_and_project_min_output_page_size | 500kB | varchar | Experimental: Minimum output page size for filter and project operators |
filter_conjunction_independence_factor | 0.75 | double | Scales the strength of independence assumption for selectivity estimates of the conjunction of multiple filters |
grouped_execution | FALSE | boolean | Use grouped execution when possible |
hash_partition_count | 100 | integer | Number of partitions for distributed joins and aggregations |
hide_inaccessible_columns | FALSE | boolean | When enabled non-accessible columns are silently filtered from results from SELECT * statements |
ignore_downstream_preferences | FALSE | boolean | Ignore Parent's PreferredProperties in AddExchange optimizer |
ignore_stats_calculator_failures | TRUE | boolean | Ignore statistics calculator failures |
incremental_hash_array_load_factor_enabled | TRUE | boolean | Use smaller load factor for small hash arrays in order to improve performance |
initial_splits_per_node | 2 | integer | The number of splits each node will run per task, initially |
iterative_optimizer_timeout | 3.00m | varchar | Timeout for plan optimization in iterative optimizer |
join_distribution_type | AUTOMATIC | varchar | Join distribution type. Possible values: [BROADCAST, PARTITIONED, AUTOMATIC] |
join_max_broadcast_table_size | 100MB | varchar | Maximum estimated size of a table that can be broadcast when using automatic join type selection |
join_multi_clause_independence_factor | 0.25 | double | Scales the strength of independence assumption for selectivity estimates of multi-clause joins |
join_reordering_strategy | AUTOMATIC | varchar | Join reordering strategy. Possible values: [NONE, ELIMINATE_CROSS_JOINS, AUTOMATIC] |
late_materialization | FALSE | boolean | Experimental: Use late materialization (including WorkProcessor pipelines) |
max_drivers_per_task | integer | Maximum number of drivers per task | |
max_partial_top_n_memory | 16MB | varchar | Max memory size for partial Top N aggregations. This can be turned off by setting it with '0'. |
max_recursion_depth | 10 | integer | Maximum recursion depth for recursive common table expression |
max_reordered_joins | 9 | integer | The maximum number of joins to reorder as one group in cost-based join reordering |
max_tasks_waiting_for_node_per_stage | 5 | integer | Maximum possible number of tasks waiting for node allocation per stage before scheduling of new tasks for stage is paused |
max_unacknowledged_splits_per_task | 500 | integer | Maximum number of leaf splits awaiting delivery to a given task |
merge_project_with_values | TRUE | boolean | Inline project expressions into values |
non_estimatable_predicate_approximation_enabled | TRUE | boolean | Approximate the cost of filters which cannot be accurately estimated even with complete statistics |
omit_datetime_type_precision | FALSE | boolean | Omit precision when rendering datetime type names with default precision |
optimize_duplicate_insensitive_joins | TRUE | boolean | Optimize duplicate insensitive joins |
optimize_hash_generation | TRUE | boolean | Compute hash codes for distribution, joins, and aggregations early in query plan |
optimize_metadata_queries | FALSE | boolean | Enable optimization for metadata queries |
optimize_mixed_distinct_aggregations | FALSE | boolean | Optimize mixed non-distinct and distinct aggregations |
optimize_top_n_ranking | TRUE | boolean | Use top N ranking optimization |
parse_decimal_literals_as_double | FALSE | boolean | Parse decimal literals as DOUBLE instead of DECIMAL |
predicate_pushdown_use_table_properties | TRUE | boolean | Use table properties in predicate pushdown |
prefer_partial_aggregation | TRUE | boolean | Prefer splitting aggregations into partial and final stages |
prefer_streaming_operators | FALSE | boolean | Prefer source table layouts that produce streaming operators |
preferred_write_partitioning_min_number_of_partitions | 50 | integer | Use preferred write partitioning when the number of written partitions exceeds the configured threshold |
push_aggregation_through_outer_join | TRUE | boolean | Allow pushing aggregations below joins |
push_partial_aggregation_through_join | FALSE | boolean | Push partial aggregations below joins |
push_table_write_through_union | TRUE | boolean | Parallelize writes when using UNION ALL in queries that write data |
query_max_cpu_time | 1000000000.00d | varchar | Maximum CPU time of a query |
query_max_execution_time | 100.00d | varchar | Maximum execution time of a query |
query_max_planning_time | 10.00m | varchar | Maximum planning time of a query |
query_max_run_time | 100.00d | varchar | Maximum run time of a query (includes the queueing time) |
query_max_scan_physical_bytes | varchar | Maximum scan physical bytes of a query | |
query_priority | 1 | integer | The priority of queries. Larger numbers are higher priority |
query_retry_attempts | 4 | integer | Maximum number of query retry attempts |
redistribute_writes | TRUE | boolean | Force parallel distributed writes |
required_workers_count | 1 | integer | Minimum number of active workers that must be available before the query will start |
required_workers_max_wait_time | 5.00m | varchar | Maximum time to wait for minimum number of workers before the query is failed |
resource_overcommit | FALSE | boolean | Use resources which are not guaranteed to be available to the query |
retry_delay_scale_factor | 2 | double | Maximum delay before initiating a retry attempt. Delay increases exponentially for each subsequent attempt starting from 'retry_initial_delay' |
retry_initial_delay | 10.00s | varchar | Initial delay before initiating a retry attempt. Delay increases exponentially for each subsequent attempt up to 'retry_max_delay' |
retry_max_delay | 1.00m | varchar | Maximum delay before initiating a retry attempt. Delay increases exponentially for each subsequent attempt starting from 'retry_initial_delay' |
rewrite_filtering_semi_join_to_inner_join | TRUE | boolean | Rewrite semi join in filtering context to inner join |
scale_writers | TRUE | boolean | Scale out writers based on throughput (use minimum necessary) |
skip_redundant_sort | TRUE | boolean | Skip redundant sort operations |
spatial_join | TRUE | boolean | Use spatial index for spatial join when possible |
spatial_partitioning_table_name | varchar | Name of the table containing spatial partitioning scheme | |
spill_enabled | FALSE | boolean | Enable spilling |
split_concurrency_adjustment_interval | 100.00ms | varchar | Experimental: Interval between changes to the number of concurrent splits per node |
statistics_cpu_timer_enabled | TRUE | boolean | Experimental: Enable cpu time tracking for automatic column statistics collection on write |
statistics_precalculation_for_pushdown_enabled | TRUE | boolean | Enable statistics precalculation for pushdown |
table_scan_node_partitioning_min_bucket_to_task_ratio | 0.5 | double | Min table scan bucket to task ratio for which plan will be adopted to node pre-partitioned tables |
task_concurrency | 1 | integer | Default number of local parallel jobs per worker |
task_retry_attempts_overall | 2147483647 | integer | Maximum number of task retry attempts overall |
task_retry_attempts_per_task | 4 | integer | Maximum number of task retry attempts per single task |
task_share_index_loading | FALSE | boolean | Share index join lookups and caching within a task |
task_writer_count | 1 | integer | Default number of local parallel table writer jobs per worker |
use_legacy_window_filter_pushdown | FALSE | boolean | Use legacy window filter pushdown optimizer |
use_mark_distinct | TRUE | boolean | Implement DISTINCT aggregations using MarkDistinct |
use_preferred_write_partitioning | TRUE | boolean | Use preferred write partitioning |
use_table_scan_node_partitioning | TRUE | boolean | Adapt plan to node pre-partitioned tables |
writer_min_size | 32MB | varchar | Target minimum size of writer output when scaling writers |
hive.bucket_execution_enabled | TRUE | boolean | Enable bucket-aware execution: only use a single worker per bucket |
hive.collect_column_statistics_on_write | TRUE | boolean | Enables automatic column level statistics collection on write |
hive.compression_codec | GZIP | varchar | Compression codec to use when writing files. Possible values: [NONE, SNAPPY, LZ4, ZSTD, GZIP] |
hive.create_empty_bucket_files | FALSE | boolean | Create empty files for buckets that have no data |
hive.dynamic_filtering_wait_timeout | 0.00m | varchar | Duration to wait for completion of dynamic filters during split generation |
hive.experimental_parquet_optimized_writer_enabled | FALSE | boolean | Experimental: Enable optimized writer |
hive.force_local_scheduling | FALSE | boolean | Only schedule splits on workers colocated with data node |
hive.hive_storage_format | ORC | varchar | Default storage format for new tables or partitions. Possible values: [ORC, PARQUET, AVRO, RCBINARY, RCTEXT, SEQUENCEFILE, JSON, TEXTFILE, CSV] |
hive.hive_views_legacy_translation | FALSE | boolean | Use legacy Hive view translation mechanism |
hive.ignore_absent_partitions | FALSE | boolean | Ignore partitions when the file system location does not exist rather than failing the query. |
hive.ignore_corrupted_statistics | FALSE | boolean | Experimental: Ignore corrupted statistics rather than failing |
hive.insert_existing_partitions_behavior | APPEND | varchar | Behavior on insert existing partitions; this session property doesn't control behavior on insert existing unpartitioned table |
hive.minimum_assigned_split_weight | 0.05 | double | Minimum assigned split weight when size based split weighting is enabled |
hive.non_transactional_optimize_enabled | FALSE | boolean | Enable OPTIMIZE table procedure |
hive.optimize_mismatched_bucket_count | FALSE | boolean | Experimental: Enable optimization to avoid shuffle when bucket count is compatible but not the same |
hive.optimize_symlink_listing | TRUE | boolean | Optimize listing for SymlinkTextFormat tables with files in a single directory |
hive.orc_bloom_filters_enabled | FALSE | boolean | ORC: Enable bloom filters for predicate pushdown |
hive.orc_lazy_read_small_ranges | TRUE | boolean | Experimental: ORC: Read small file segments lazily |
hive.orc_max_buffer_size | 8MB | varchar | ORC: Maximum size of a single read |
hive.orc_max_merge_distance | 1MB | varchar | ORC: Maximum size of gap between two reads to merge into a single read |
hive.orc_max_read_block_size | 16MB | varchar | ORC: Soft max size of Trino blocks produced by ORC reader |
hive.orc_nested_lazy_enabled | TRUE | boolean | Experimental: ORC: Lazily read nested data |
hive.orc_optimized_writer_max_dictionary_memory | 16MB | varchar | ORC: Max dictionary memory |
hive.orc_optimized_writer_max_stripe_rows | 10000000 | integer | ORC: Max stripe row count |
hive.orc_optimized_writer_max_stripe_size | 64MB | varchar | ORC: Max stripe size |
hive.orc_optimized_writer_min_stripe_size | 32MB | varchar | ORC: Min stripe size |
hive.orc_optimized_writer_validate | FALSE | boolean | ORC: Force all validation for files |
hive.orc_optimized_writer_validate_mode | BOTH | varchar | ORC: Level of detail in ORC validation. Possible values: [HASHED, DETAILED, BOTH] |
hive.orc_optimized_writer_validate_percentage | 0 | double | ORC: sample percentage for validation for files |
hive.orc_stream_buffer_size | 8MB | varchar | ORC: Size of buffer for streaming reads |
hive.orc_string_statistics_limit | 64B | varchar | ORC: Maximum size of string statistics; drop if exceeding |
hive.orc_tiny_stripe_threshold | 8MB | varchar | ORC: Threshold below which an ORC stripe or file will read in its entirety |
hive.orc_use_column_names | FALSE | boolean | ORC: Access ORC columns using names from the file |
hive.parallel_partitioned_bucketed_writes | TRUE | boolean | Improve parallelism of partitioned and bucketed table writes |
hive.parquet_ignore_statistics | FALSE | boolean | Ignore statistics from Parquet to allow querying files with corrupted or incorrect statistics |
hive.parquet_max_read_block_size | 16MB | varchar | Parquet: Maximum size of a block to read |
hive.parquet_use_column_index | TRUE | boolean | Use Parquet column index |
hive.parquet_use_column_names | TRUE | boolean | Parquet: Access Parquet columns using names from the file |
hive.parquet_writer_batch_size | 10000 | integer | Parquet: Maximum number of rows passed to the writer in each batch |
hive.parquet_writer_block_size | 134217728B | varchar | Parquet: Writer block size |
hive.parquet_writer_page_size | 1048576B | varchar | Parquet: Writer page size |
hive.partition_statistics_sample_size | 100 | integer | Maximum sample size of the partitions column statistics |
hive.projection_pushdown_enabled | TRUE | boolean | Projection push down enabled for hive |
hive.propagate_table_scan_sorting_properties | FALSE | boolean | Use sorted table layout to generate more efficient execution plans. May lead to incorrect results if files are not sorted as per table definition. |
hive.query_partition_filter_required | FALSE | boolean | Require filter on partition column |
hive.query_partition_filter_required_schemas | [] | array(varchar) | List of schemas for which filter on partition column is enforced. |
hive.rcfile_optimized_writer_validate | FALSE | boolean | RCFile: Validate writer files |
hive.respect_table_format | TRUE | boolean | Write new partitions using table format rather than default storage format |
hive.s3_select_pushdown_enabled | FALSE | boolean | S3 Select pushdown enabled |
hive.size_based_split_weights_enabled | TRUE | boolean | Enable estimating split weights based on size in bytes |
hive.sorted_writing_enabled | TRUE | boolean | Enable writing to bucketed sorted tables |
hive.statistics_enabled | TRUE | boolean | Expose table statistics |
hive.target_max_file_size | 1GB | varchar | Target maximum size of written files; the actual size may be larger |
hive.temporary_staging_directory_enabled | TRUE | boolean | Should use temporary staging directory for write operations |
hive.temporary_staging_directory_path | /tmp/presto-${USER} | varchar | Temporary staging directory location |
hive.timestamp_precision | MILLISECONDS | varchar | Precision for timestamp columns in Hive tables. Possible values: [MILLISECONDS, MICROSECONDS, NANOSECONDS] |
hive.validate_bucketing | TRUE | boolean | Verify that data is bucketed correctly when reading |
mysql.aggregation_pushdown_enabled | TRUE | boolean | Enable aggregation pushdown |
mysql.decimal_default_scale | 0 | integer | Default decimal scale for mapping unspecified and exceeding precision decimals. Not used when decimal_mapping is set to STRICT |
mysql.decimal_mapping | STRICT | varchar | Decimal mapping for unspecified and exceeding precision decimals. STRICT skips them. ALLOW_OVERFLOW requires setting proper decimal scale and rounding mode. Possible values: [STRICT, ALLOW_OVERFLOW] |
mysql.decimal_rounding_mode | UNNECESSARY | varchar | Rounding mode for mapping unspecified and exceeding precision decimals. Not used when decimal_mapping is set to STRICT. Possible values: [UP, DOWN, CEILING, FLOOR, HALF_UP, HALF_DOWN, HALF_EVEN, UNNECESSARY] |
mysql.domain_compaction_threshold | 32 | integer | Maximum ranges to allow in a tuple domain without simplifying it |
mysql.join_pushdown_automatic_max_join_to_tables_ratio | 1.25 | double | If estimated join output size is greater than or equal to ratio * sum of table sizes, then join pushdown will not be performed |
mysql.join_pushdown_automatic_max_table_size | varchar | Maximum table size to be considered for join pushdown | |
mysql.join_pushdown_enabled | TRUE | boolean | Enable join pushdown |
mysql.join_pushdown_strategy | AUTOMATIC | varchar | Join pushdown strategy. Possible values: [EAGER, AUTOMATIC] |
mysql.non_transactional_insert | FALSE | boolean | Do not use temporary table on insert to table |
mysql.topn_pushdown_enabled | TRUE | boolean | Enable TopN pushdown |
mysql.unsupported_type_handling | IGNORE | varchar | Unsupported type handling strategy. Possible values: [IGNORE, CONVERT_TO_VARCHAR] |
mysql.write_batch_size | 1000 | integer | Maximum number of rows to write in a single batch |
from dbt-trino.
Related Issues (20)
- Include query_id in run results for seed HOT 1
- Invalid SQL generated for dbt_project_evaluator HOT 1
- dbt-trino adds comment into table create statement by default HOT 2
- dbt-Trino snapshot cannot create __dbt_tmp table after first run HOT 12
- Add tests for aliases
- Add copyright notices to files HOT 1
- Support elementary in improving data-quality capabilities HOT 3
- Extend Hive test coverage
- upgrade to support dbt-core v1.7.0 HOT 1
- Incorrect Schema Used When Renaming Materialized Views HOT 4
- Support CASCADE dropping relations
- Solving for large stage depths HOT 3
- Support `CREATE OR REPLACE` HOT 2
- get_relation not working as already_exists HOT 4
- deltalake rename managed table not allowed arised HOT 1
- Failed to connect to Trino cluster using LDAP auth and HTTP connection HOT 3
- Extra credentials in connection HOT 3
- Not able to create snapshot of a model using dbt-trino==1.7.1 HOT 4
- Refactor to use dbt-adapters interface layer HOT 2
- merge_exclude_columns doesn't work
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from dbt-trino.