From 466355f71a77d6803fe96b260c813ed7c4c1693c Mon Sep 17 00:00:00 2001 From: Gabor Szarnyas Date: Sun, 8 Dec 2024 21:47:11 +0100 Subject: [PATCH 1/4] Update Contributor Guide on table types --- CONTRIBUTING.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 6999f17ed34..245c58caa25 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -70,7 +70,8 @@ Some of this style guide is automated with GitHub Actions, but feel free to run * Always format SQL code, variable names, function names, etc. as code. For example, when talking about the `CREATE TABLE` statement, the keywords should be formatted as code. * When presenting SQL statements, do not include the DuckDB prompt (`D `). * SQL statements should end with a semicolon (`;`) to allow readers to quickly paste them into a SQL console. -* Narrow tables – that do not span horizontally across the entire page – should be prepended with an empty div that has the `narrow_table` class: `
`. +* Tables with predominantly code output (e.g., the result of a `DESCRIBE` statement) should be prepended with an empty div that has the `monospace_table` class: `
`. +* Tables where the headers should be center-aligned (opposed to the left-aligned default) should be prepended with an empty div that has the `center_aligned_header_table` class: `
`. * Do not introduce hard line breaks if possible. Therefore, avoid using the `
` HTML tag and avoid [double spaces at the end of a line in Markdown](https://spec.commonmark.org/0.28/#hard-line-breaks). * Single and double quote characters (`'` and `"`) are not converted to smart quotation marks automatically. To insert these, use `“` `”` and `‘` `’`. * When referencing other articles, put their titles in quotes, e.g., `see the [“Lightweight Compression in DuckDB” blog post]({% post_url 2022-10-28-lightweight-compression %})`. From 5ce2efa19a17ffa6fd582b4672eaec2886b30716 Mon Sep 17 00:00:00 2001 From: Gabor Szarnyas Date: Sun, 8 Dec 2024 21:57:23 +0100 Subject: [PATCH 2/4] Remove narrow_table divs --- _posts/2022-05-04-friendlier-sql.md | 5 ---- _posts/2023-07-07-python-udf.md | 2 -- _posts/2023-08-23-even-friendlier-sql.md | 23 ------------------- ...09-15-asof-joins-fuzzy-temporal-lookups.md | 15 ------------ _posts/2023-09-26-announcing-duckdb-090.md | 3 --- _posts/2023-10-27-csv-sniffer.md | 1 - _posts/2024-02-13-announcing-duckdb-0100.md | 3 --- _posts/2024-07-05-community-extensions.md | 1 - docs/api/adbc.md | 1 - docs/api/cli/arguments.md | 1 - docs/api/cli/dot_commands.md | 1 - docs/api/cli/editing.md | 5 ---- docs/api/cli/output_formats.md | 1 - docs/api/cli/syntax_highlighting.md | 2 -- docs/archive/0.10/api/adbc.md | 1 - docs/archive/0.10/api/cli/arguments.md | 1 - docs/archive/0.10/api/cli/dot_commands.md | 1 - docs/archive/0.10/api/cli/editing.md | 5 ---- docs/archive/0.10/api/cli/output_formats.md | 1 - .../0.10/api/cli/syntax_highlighting.md | 2 -- docs/archive/0.10/api/python/types.md | 3 --- docs/archive/0.10/configuration/pragmas.md | 1 - docs/archive/0.10/data/csv/auto_detection.md | 4 ---- docs/archive/0.10/data/csv/overview.md | 3 --- .../0.10/data/csv/reading_faulty_csv_files.md | 7 ------ docs/archive/0.10/data/json/overview.md | 9 -------- .../data/multiple_files/combining_schemas.md | 2 -- .../0.10/data/multiple_files/overview.md | 3 --- docs/archive/0.10/data/parquet/metadata.md | 4 ---- docs/archive/0.10/dev/release_calendar.md | 2 -- docs/archive/0.10/extensions/arrow.md | 1 - docs/archive/0.10/extensions/autocomplete.md | 2 -- .../0.10/extensions/full_text_search.md | 1 - docs/archive/0.10/extensions/httpfs/s3api.md | 3 --- .../httpfs/s3api_legacy_authentication.md | 1 - docs/archive/0.10/extensions/json.md | 8 ------- docs/archive/0.10/extensions/mysql.md | 1 - .../0.10/extensions/official_extensions.md | 1 - docs/archive/0.10/extensions/overview.md | 1 - docs/archive/0.10/extensions/spatial.md | 1 - docs/archive/0.10/extensions/tpch.md | 1 - .../extensions/versioning_of_extensions.md | 4 ---- .../extensions/working_with_extensions.md | 2 -- .../0.10/guides/database_integration/mysql.md | 1 - .../guides/database_integration/postgres.md | 1 - docs/archive/0.10/guides/meta/describe.md | 2 -- docs/archive/0.10/guides/meta/list_tables.md | 2 -- .../network_cloud_storage/s3_express_one.md | 1 - .../0.10/guides/performance/file_formats.md | 2 -- .../0.10/guides/performance/indexing.md | 1 - .../archive/0.10/guides/performance/schema.md | 3 --- .../0.10/guides/sql_features/asof_join.md | 4 ---- docs/archive/0.10/internals/storage.md | 1 - docs/archive/0.10/sql/data_types/bitstring.md | 1 - docs/archive/0.10/sql/data_types/blob.md | 1 - docs/archive/0.10/sql/data_types/boolean.md | 3 --- docs/archive/0.10/sql/data_types/date.md | 2 -- docs/archive/0.10/sql/data_types/enum.md | 2 -- docs/archive/0.10/sql/data_types/interval.md | 1 - docs/archive/0.10/sql/data_types/list.md | 1 - docs/archive/0.10/sql/data_types/numeric.md | 2 -- docs/archive/0.10/sql/data_types/struct.md | 1 - docs/archive/0.10/sql/data_types/text.md | 1 - docs/archive/0.10/sql/data_types/time.md | 1 - docs/archive/0.10/sql/data_types/timestamp.md | 2 -- docs/archive/0.10/sql/data_types/timezones.md | 1 - docs/archive/0.10/sql/data_types/union.md | 1 - .../0.10/sql/duckdb_table_functions.md | 4 ---- .../sql/expressions/comparison_operators.md | 3 --- .../0.10/sql/expressions/logical_operators.md | 2 -- docs/archive/0.10/sql/expressions/star.md | 8 ------- .../0.10/sql/expressions/subqueries.md | 1 - docs/archive/0.10/sql/functions/bitstring.md | 1 - docs/archive/0.10/sql/functions/char.md | 2 -- docs/archive/0.10/sql/functions/dateformat.md | 1 - docs/archive/0.10/sql/functions/numeric.md | 1 - .../0.10/sql/functions/pattern_matching.md | 4 ---- .../0.10/sql/functions/regular_expressions.md | 2 -- docs/archive/0.10/sql/functions/time.md | 1 - docs/archive/0.10/sql/functions/utility.md | 1 - docs/archive/0.10/sql/query_syntax/filter.md | 5 ---- docs/archive/0.10/sql/query_syntax/orderby.md | 2 -- docs/archive/0.10/sql/query_syntax/qualify.md | 1 - .../0.10/sql/statements/create_index.md | 2 -- docs/archive/0.10/sql/statements/insert.md | 3 --- docs/archive/0.10/sql/statements/pivot.md | 12 ---------- docs/archive/0.10/sql/statements/unpivot.md | 7 ------ docs/archive/0.10/sql/statements/update.md | 2 -- docs/archive/0.10/sql/window_functions.md | 3 --- docs/archive/0.9/api/adbc.md | 1 - docs/archive/0.9/api/cli.md | 3 --- docs/archive/0.9/api/python/types.md | 3 --- docs/archive/0.9/data/csv/auto_detection.md | 4 ---- docs/archive/0.9/data/csv/overview.md | 3 --- docs/archive/0.9/data/json/overview.md | 9 -------- .../data/multiple_files/combining_schemas.md | 2 -- .../0.9/data/multiple_files/overview.md | 3 --- docs/archive/0.9/data/parquet/metadata.md | 2 -- docs/archive/0.9/extensions/autocomplete.md | 2 -- .../0.9/extensions/full_text_search.md | 1 - docs/archive/0.9/extensions/httpfs.md | 4 ---- docs/archive/0.9/extensions/mysql.md | 1 - docs/archive/0.9/extensions/spatial.md | 1 - docs/archive/0.9/guides/meta/list_tables.md | 3 --- docs/archive/0.9/sql/data_types/bitstring.md | 1 - docs/archive/0.9/sql/data_types/blob.md | 1 - docs/archive/0.9/sql/data_types/boolean.md | 3 --- docs/archive/0.9/sql/data_types/date.md | 3 --- docs/archive/0.9/sql/data_types/enum.md | 2 -- docs/archive/0.9/sql/data_types/interval.md | 1 - docs/archive/0.9/sql/data_types/numeric.md | 2 -- docs/archive/0.9/sql/data_types/struct.md | 4 ---- docs/archive/0.9/sql/data_types/text.md | 1 - docs/archive/0.9/sql/data_types/timestamp.md | 2 -- docs/archive/0.9/sql/data_types/timezones.md | 1 - docs/archive/0.9/sql/data_types/union.md | 1 - .../archive/0.9/sql/duckdb_table_functions.md | 3 --- .../sql/expressions/comparison_operators.md | 3 --- .../0.9/sql/expressions/logical_operators.md | 2 -- docs/archive/0.9/sql/expressions/star.md | 6 ----- .../archive/0.9/sql/expressions/subqueries.md | 1 - docs/archive/0.9/sql/functions/bitstring.md | 1 - docs/archive/0.9/sql/functions/dateformat.md | 1 - docs/archive/0.9/sql/functions/numeric.md | 1 - .../0.9/sql/functions/patternmatching.md | 6 ----- docs/archive/0.9/sql/functions/time.md | 1 - docs/archive/0.9/sql/functions/utility.md | 1 - docs/archive/0.9/sql/indexes.md | 2 -- docs/archive/0.9/sql/pragmas.md | 1 - docs/archive/0.9/sql/query_syntax/filter.md | 5 ---- docs/archive/0.9/sql/query_syntax/orderby.md | 2 -- docs/archive/0.9/sql/query_syntax/qualify.md | 1 - docs/archive/0.9/sql/statements/insert.md | 3 --- docs/archive/0.9/sql/statements/pivot.md | 12 ---------- docs/archive/0.9/sql/statements/unpivot.md | 7 ------ docs/archive/0.9/sql/statements/update.md | 2 -- docs/archive/1.0/api/adbc.md | 1 - docs/archive/1.0/api/cli/arguments.md | 1 - docs/archive/1.0/api/cli/dot_commands.md | 1 - docs/archive/1.0/api/cli/editing.md | 5 ---- docs/archive/1.0/api/cli/output_formats.md | 1 - .../1.0/api/cli/syntax_highlighting.md | 2 -- docs/archive/1.0/configuration/pragmas.md | 1 - docs/archive/1.0/data/csv/auto_detection.md | 1 - docs/archive/1.0/data/csv/overview.md | 3 --- .../1.0/data/csv/reading_faulty_csv_files.md | 5 ---- .../data/multiple_files/combining_schemas.md | 2 -- .../1.0/data/multiple_files/overview.md | 3 --- docs/archive/1.0/dev/release_calendar.md | 2 -- docs/archive/1.0/extensions/arrow.md | 1 - docs/archive/1.0/extensions/autocomplete.md | 2 -- .../archive/1.0/extensions/core_extensions.md | 1 - .../1.0/extensions/full_text_search.md | 1 - docs/archive/1.0/extensions/httpfs/s3api.md | 3 --- .../httpfs/s3api_legacy_authentication.md | 1 - docs/archive/1.0/extensions/json.md | 3 --- docs/archive/1.0/extensions/overview.md | 1 - docs/archive/1.0/extensions/spatial.md | 1 - docs/archive/1.0/extensions/tpch.md | 1 - .../extensions/versioning_of_extensions.md | 4 ---- .../1.0/extensions/working_with_extensions.md | 2 -- .../1.0/guides/database_integration/mysql.md | 1 - .../guides/database_integration/postgres.md | 1 - docs/archive/1.0/guides/meta/describe.md | 2 -- docs/archive/1.0/guides/meta/list_tables.md | 2 -- .../network_cloud_storage/s3_express_one.md | 1 - .../1.0/guides/performance/file_formats.md | 2 -- .../1.0/guides/performance/indexing.md | 1 - docs/archive/1.0/guides/performance/schema.md | 3 --- .../1.0/guides/sql_features/asof_join.md | 4 ---- docs/archive/1.0/internals/storage.md | 1 - docs/archive/1.0/sql/data_types/bitstring.md | 1 - docs/archive/1.0/sql/data_types/blob.md | 1 - docs/archive/1.0/sql/data_types/boolean.md | 3 --- docs/archive/1.0/sql/data_types/date.md | 2 -- docs/archive/1.0/sql/data_types/enum.md | 2 -- docs/archive/1.0/sql/data_types/interval.md | 1 - docs/archive/1.0/sql/data_types/numeric.md | 2 -- docs/archive/1.0/sql/data_types/struct.md | 1 - docs/archive/1.0/sql/data_types/text.md | 1 - docs/archive/1.0/sql/data_types/time.md | 1 - docs/archive/1.0/sql/data_types/timestamp.md | 2 -- docs/archive/1.0/sql/data_types/timezones.md | 1 - docs/archive/1.0/sql/data_types/union.md | 1 - .../sql/expressions/comparison_operators.md | 3 --- .../1.0/sql/expressions/logical_operators.md | 2 -- docs/archive/1.0/sql/expressions/star.md | 8 ------- .../archive/1.0/sql/expressions/subqueries.md | 1 - docs/archive/1.0/sql/functions/bitstring.md | 1 - docs/archive/1.0/sql/functions/char.md | 2 -- docs/archive/1.0/sql/functions/dateformat.md | 1 - docs/archive/1.0/sql/functions/numeric.md | 1 - .../1.0/sql/functions/pattern_matching.md | 1 - .../1.0/sql/functions/regular_expressions.md | 2 -- docs/archive/1.0/sql/functions/time.md | 1 - docs/archive/1.0/sql/functions/utility.md | 1 - .../1.0/sql/functions/window_functions.md | 3 --- .../1.0/sql/meta/duckdb_table_functions.md | 4 ---- docs/archive/1.0/sql/query_syntax/orderby.md | 2 -- docs/archive/1.0/sql/query_syntax/qualify.md | 1 - .../1.0/sql/statements/create_index.md | 2 -- docs/archive/1.0/sql/statements/insert.md | 3 --- docs/archive/1.0/sql/statements/pivot.md | 12 ---------- docs/archive/1.0/sql/statements/unpivot.md | 7 ------ docs/archive/1.0/sql/statements/update.md | 2 -- docs/configuration/pragmas.md | 1 - docs/data/csv/auto_detection.md | 1 - docs/data/csv/overview.md | 3 --- docs/data/csv/reading_faulty_csv_files.md | 5 ---- docs/data/json/creating_json.md | 1 - docs/data/json/json_functions.md | 2 -- docs/data/multiple_files/combining_schemas.md | 2 -- docs/data/multiple_files/overview.md | 3 --- docs/dev/release_calendar.md | 2 -- docs/extensions/arrow.md | 1 - docs/extensions/autocomplete.md | 2 -- docs/extensions/core_extensions.md | 1 - docs/extensions/full_text_search.md | 1 - docs/extensions/httpfs/s3api.md | 3 --- .../httpfs/s3api_legacy_authentication.md | 1 - docs/extensions/overview.md | 1 - docs/extensions/tpch.md | 1 - docs/extensions/versioning_of_extensions.md | 2 -- docs/extensions/working_with_extensions.md | 1 - docs/guides/database_integration/mysql.md | 1 - docs/guides/database_integration/postgres.md | 1 - docs/guides/meta/describe.md | 2 -- docs/guides/meta/list_tables.md | 2 -- .../network_cloud_storage/s3_express_one.md | 1 - docs/guides/performance/file_formats.md | 2 -- docs/guides/performance/indexing.md | 1 - docs/guides/performance/schema.md | 3 --- docs/guides/sql_features/asof_join.md | 4 ---- docs/internals/pivot.md | 3 --- docs/internals/storage.md | 1 - docs/sql/data_types/bitstring.md | 1 - docs/sql/data_types/blob.md | 1 - docs/sql/data_types/boolean.md | 3 --- docs/sql/data_types/date.md | 2 -- docs/sql/data_types/enum.md | 2 -- docs/sql/data_types/interval.md | 1 - docs/sql/data_types/numeric.md | 2 -- docs/sql/data_types/struct.md | 1 - docs/sql/data_types/text.md | 1 - docs/sql/data_types/time.md | 1 - docs/sql/data_types/timestamp.md | 2 -- docs/sql/data_types/timezones.md | 1 - docs/sql/data_types/union.md | 1 - docs/sql/dialect/postgresql_compatibility.md | 4 ---- docs/sql/expressions/comparison_operators.md | 3 --- docs/sql/expressions/logical_operators.md | 2 -- docs/sql/expressions/star.md | 8 ------- docs/sql/expressions/subqueries.md | 1 - docs/sql/functions/bitstring.md | 1 - docs/sql/functions/char.md | 2 -- docs/sql/functions/dateformat.md | 1 - docs/sql/functions/numeric.md | 1 - docs/sql/functions/pattern_matching.md | 1 - docs/sql/functions/regular_expressions.md | 2 -- docs/sql/functions/time.md | 1 - docs/sql/functions/utility.md | 1 - docs/sql/functions/window_functions.md | 3 --- docs/sql/meta/duckdb_table_functions.md | 4 ---- docs/sql/query_syntax/orderby.md | 2 -- docs/sql/query_syntax/qualify.md | 1 - docs/sql/statements/attach.md | 1 - docs/sql/statements/create_index.md | 2 -- docs/sql/statements/insert.md | 3 --- docs/sql/statements/pivot.md | 10 -------- docs/sql/statements/unpivot.md | 6 ----- docs/sql/statements/update.md | 2 -- 271 files changed, 653 deletions(-) diff --git a/_posts/2022-05-04-friendlier-sql.md b/_posts/2022-05-04-friendlier-sql.md index 24b9eb7450c..54127c98e91 100644 --- a/_posts/2022-05-04-friendlier-sql.md +++ b/_posts/2022-05-04-friendlier-sql.md @@ -118,7 +118,6 @@ CREATE TABLE mandalorian AS SELECT 1 AS "THIS_IS_THE_WAY"; SELECT this_is_the_way FROM mandalorian; ``` -
| THIS_IS_THE_WAY | |----------------:| @@ -163,7 +162,6 @@ Even as SQL fans, we know that SQL can learn a thing or two from newer languages SELECT 'I love you! I know'[:-3] AS nearly_soloed; ``` -
| nearly_soloed | |:---| @@ -189,7 +187,6 @@ SELECT FROM (SELECT ['A-Wing', 'B-Wing', 'X-Wing', 'Y-Wing'] AS starfighter_list); ``` -
| dont_forget_the_b_wing | |:---| @@ -249,7 +246,6 @@ FROM ( ) theyre_coming_in_too_fast; ``` -
| tie_fighter | tie_fighter:1 | |:---|:---| @@ -270,7 +266,6 @@ JOIN sith_count_varchar s_char ON s_int.sith_count = s_char.sith_count; ``` -
| sith_count | sith_count | |---:|---:| diff --git a/_posts/2023-07-07-python-udf.md b/_posts/2023-07-07-python-udf.md index 47b782e2785..21a45deeb1c 100644 --- a/_posts/2023-07-07-python-udf.md +++ b/_posts/2023-07-07-python-udf.md @@ -236,7 +236,6 @@ native_res = con.sql("SELECT sum(add_built_in_type(i)) FROM numbers").fetchall() arrow_res = con.sql("SELECT sum(add_arrow_type(i)) FROM numbers").fetchall() ``` -
| Name | Time (s) | |-------------|---------:| @@ -295,7 +294,6 @@ con.sql("SELECT sum(strlen_arrow(i)) FROM strings tbl(i)").fetchall() exec_external(con) ``` -
| Name | Time (s) | Peak memory consumption (MB) | |-------------|---------:|-----------------------------:| diff --git a/_posts/2023-08-23-even-friendlier-sql.md b/_posts/2023-08-23-even-friendlier-sql.md index 8c76eb02622..6b50ab582a3 100644 --- a/_posts/2023-08-23-even-friendlier-sql.md +++ b/_posts/2023-08-23-even-friendlier-sql.md @@ -65,7 +65,6 @@ SELECT substr(intro, starship_loc + len('starship') + 1) AS trimmed_intro; ``` -
| intro | starship_loc | trimmed_intro | |:---|:---|:---| @@ -90,7 +89,6 @@ CREATE TABLE trek_facts AS DESCRIBE trek_facts; ``` -
| column_name | column_type | null | key | default | extra | |:---|:---|:---|:---|:---|:---| @@ -124,7 +122,6 @@ SELECT FROM trek_facts; ``` -
| episode_num | cnt_warp_speed_orders | highest_warp_speed_issued | |:---|:---|:---| @@ -145,7 +142,6 @@ SELECT FROM trek_facts; ``` -
| max(trek_facts.cnt_warp_speed_orders) | max(trek_facts.highest_warp_speed_issued) | |:---|:---| @@ -165,7 +161,6 @@ WHERE -- highest_warp_speed_issued >= 2 ``` -
| episode_num | cnt_warp_speed_orders | highest_warp_speed_issued | |:---|:---|:---| @@ -184,7 +179,6 @@ SELECT FROM trek_facts; ``` -
| max(trek_facts.
episode_num) | max(trek_facts.
aired_date) | max(trek_facts.
cnt_kirk_hookups) | ... | max(trek_facts.
bool_enterprise_saved_the_day) | |:---|:---|:---|:---|:---| @@ -198,7 +192,6 @@ SELECT FROM trek_facts; ``` -
| max(trek_facts.
season_num) | max(trek_facts.
episode_num) | max(aired_date :=
CAST(aired_date AS TIMESTAMP)) | ... | max(trek_facts.
bool_enterprise_saved_the_day) | |:---|:---|:---|:---|:---| @@ -219,7 +212,6 @@ WHERE COLUMNS(col -> col LIKE '%warp%') >= 2; ``` -
| episode_num | cnt_warp_speed_orders | highest_warp_speed_issued | |:---|:---|:---| @@ -245,7 +237,6 @@ SELECT FROM 'https://raw.githubusercontent.com/vlad-saling/star-trek-ipsum/master/src/content/content.json'; ``` -
| starship | |:---| @@ -293,7 +284,6 @@ SELECT .concat('.') AS im_not_messing_around_number_one; ``` -
| im_not_messing_around_number_one | |:---| @@ -312,7 +302,6 @@ SELECT '.') AS oof; ``` -
| oof | |:---| @@ -336,7 +325,6 @@ CREATE TABLE proverbs AS FROM proverbs; ``` -
| klingon_proverb | borg_proverb | |:---|:---| @@ -358,7 +346,6 @@ INSERT INTO proverbs BY NAME SELECT * FROM proverbs; ``` -
| klingon_proverb | borg_proverb | |:---|:---| @@ -386,7 +373,6 @@ INSERT INTO purchases FROM purchases; ``` -
| item | year | count | |:---|:---|:---| @@ -409,7 +395,6 @@ CREATE TABLE pivoted_purchases AS FROM pivoted_purchases; ``` -
| item | 2155 | 2156 | 2157 | |:---|:---|:---|:---| @@ -430,7 +415,6 @@ UNPIVOT pivoted_purchases VALUE count; ``` -
| item | year | count | |:---|:---|:---| @@ -457,7 +441,6 @@ SELECT .list_transform(x -> x.string_split(' ')[1]) AS short_name; ``` -
| ship_name | |:---| @@ -471,7 +454,6 @@ SELECT .list_filter(x -> x.contains('1701')) AS the_original; ``` -
| the_original | |:---| @@ -490,7 +472,6 @@ SELECT IF x.contains('1701')] AS ready_to_boldly_go; ``` -
| ready_to_boldly_go | |:---| @@ -509,7 +490,6 @@ SELECT casualties.*; ``` -
| gold_casualties | blue_casualties | red_casualties | |:---|:---|:---| @@ -529,7 +509,6 @@ FROM officers SELECT officers; ``` -
| officers | |:---| @@ -554,7 +533,6 @@ SELECT 6 UNION ALL SELECT 'First Contact'; ``` -
| movie | |:---| @@ -582,7 +560,6 @@ SELECT movie.num; ``` -
| movie | type | name | num | |:---|:---|:---|:---| diff --git a/_posts/2023-09-15-asof-joins-fuzzy-temporal-lookups.md b/_posts/2023-09-15-asof-joins-fuzzy-temporal-lookups.md index 489ddfccccc..144a0af9a3f 100644 --- a/_posts/2023-09-15-asof-joins-fuzzy-temporal-lookups.md +++ b/_posts/2023-09-15-asof-joins-fuzzy-temporal-lookups.md @@ -35,7 +35,6 @@ which can be cumbersome and slow to implement in standard SQL. Let's start with a concrete example. Suppose we have a table of stock [`prices`](/data/prices.csv) with timestamps: -
| ticker | when | price | | :----- | :--- | ----: | @@ -51,7 +50,6 @@ Suppose we have a table of stock [`prices`](/data/prices.csv) with timestamps: We have another table containing portfolio [`holdings`](/data/holdings.csv) at various points in time: -
| ticker | when | shares | | :----- | :--- | -----: | @@ -77,7 +75,6 @@ FROM holdings h ASOF JOIN prices p This attaches the value of the holding at that time to each row: -
| ticker | when | value | | :----- | :--- | ----: | @@ -107,7 +104,6 @@ ORDER BY ALL; As you might expect, this will produce `NULL` prices and values instead of dropping left side rows when there is no ticker or the time is before the prices begin. -
| ticker | when | value | | :----- | :--- | ----: | @@ -148,7 +144,6 @@ INNER JOIN state s The default value of `infinity` is used to make sure there is an end value for the last row that can be compared. Here is what the `state` CTE looks like for our example: -
| ticker | price | when | end | |:-------|------:|:--------------------|:--------------------| @@ -227,7 +222,6 @@ But AsOf can now use any inequality, which allows it to handle other types of ev To explore this, let's use two very simple tables with no equality conditions. The build side will just have four integer "timestamps" with alphabetic values: -
| Time | Value | | ---: | ----: | @@ -240,7 +234,6 @@ The probe table will just be the time values plus the midpoints, and we can make a table showing what value each probe time matches for greater than or equal to: -
| Probe | >= | | ----: | --- | @@ -258,7 +251,6 @@ This shows us that the interval a probe value matches is in the half-open interv Now let's see what happens if use strictly greater than as the inequality: -
| Probe | > | | ----: | --- | @@ -278,7 +270,6 @@ This means that for this inequality type, the time is not part of the interval. What if the inequality goes in the other direction, say less than or equal to? -
| Probe | <= | | ----: | --- | @@ -303,7 +294,6 @@ when non-strict inequalities are used. We can check this by looking at the last inequality: strictly less than: -
| Probe | < | | ----: | --- | @@ -323,7 +313,6 @@ and it is a less than, so the time is the end of the interval. To sum up, here is the full list: -
| Inequality | Interval | | -- | ---------- | @@ -445,7 +434,6 @@ CREATE OR REPLACE TABLE probe AS ( The `build` table looks like this: -
| k | t | v | |---|---------------------|---| @@ -457,7 +445,6 @@ The `build` table looks like this: and the probe table looks like this (with only even values for k): -
| k | t | |---|---------------------| @@ -501,7 +488,6 @@ than inequalities and generates a hash join with a filter. Running the benchmark, we get results like this: -
| Algorithm | Median of 5 | | :--------- | ----------: | @@ -551,7 +537,6 @@ INNER JOIN state s AND p.k = s.k; ``` -
| Algorithm | Median of 5 runs | | :--------- | ---------------: | diff --git a/_posts/2023-09-26-announcing-duckdb-090.md b/_posts/2023-09-26-announcing-duckdb-090.md index 917aa3c5cee..3e8b97fc922 100644 --- a/_posts/2023-09-26-announcing-duckdb-090.md +++ b/_posts/2023-09-26-announcing-duckdb-090.md @@ -66,7 +66,6 @@ SELECT count(*) FROM (SELECT DISTINCT * FROM tbl); If we keep all the data in memory, the query should use around 6GB. However, we can still complete the query if less memory is available. In the table below, we can see how the runtime is affected by lowering the memory limit: -
| memory limit | v0.8.1 | v0.9.0 | |--------------:|---------:|---------:| @@ -135,7 +134,6 @@ SELECT FROM tripdata; ``` -
| Version | Run time | |--------:|---------:| @@ -153,7 +151,6 @@ CREATE TABLE integers(i INTEGER PRIMARY KEY); INSERT INTO integers FROM range(10000000); ``` -
| Version | Size | | -- | --: | diff --git a/_posts/2023-10-27-csv-sniffer.md b/_posts/2023-10-27-csv-sniffer.md index c54293959e7..385329f044a 100644 --- a/_posts/2023-10-27-csv-sniffer.md +++ b/_posts/2023-10-27-csv-sniffer.md @@ -174,7 +174,6 @@ To analyze the impact of running DuckDB's automatic detection, we execute the sn The cost of sniffing the dialect column names and types is approximately 4% of the total cost of loading the data. -
| Name | Time (s) | |-------------|----------| diff --git a/_posts/2024-02-13-announcing-duckdb-0100.md b/_posts/2024-02-13-announcing-duckdb-0100.md index 9299d154a3b..dbb65b21a9a 100644 --- a/_posts/2024-02-13-announcing-duckdb-0100.md +++ b/_posts/2024-02-13-announcing-duckdb-0100.md @@ -187,7 +187,6 @@ We expect that as the format stabilizes and matures this will happen less freque Below is a benchmark comparing the loading time of 11 million rows of the NYC Taxi dataset from a CSV file on an M1 Max with 10 cores: -
| Version | Load time | |----------|-----------:| @@ -196,7 +195,6 @@ Below is a benchmark comparing the loading time of 11 million rows of the NYC Ta Furthermore, many optimizations have been done that make running queries over CSV files directly significantly faster as well. Below is a benchmark comparing the execution time of a `SELECT count(*)` query directly over the NYC Taxi CSV file. -
| Version | Query time | |----------|-----------:| @@ -335,7 +333,6 @@ With the new version 0.10.0, this query completes in ca. 5s on a MacBook, while Floating point numbers are notoriously difficult to compress efficiently, both in terms of compression ratio as well as speed of compression and decompression. In the past, DuckDB had support for the then state-of-the-art "[Chimp](https://github.com/duckdb/duckdb/pull/4878)" and the "[Patas](https://github.com/duckdb/duckdb/pull/5044)" compression methods. Turns out, those were not the last word in floating point compression. Researchers [Azim Afroozeh](https://www.cwi.nl/en/people/azim-afroozeh/), [Leonard Kuffo](https://www.cwi.nl/en/people/leonardo-xavier-kuffo-rivero/) and (the one and only) [Peter Boncz](https://homepages.cwi.nl/~boncz/) have recently published a paper titled "[ALP: Adaptive Lossless floating-Point Compression](https://dl.acm.org/doi/pdf/10.1145/3626717)" at SIGMOD, a top-tier academic conference for data management research. In an uncommon yet highly commendable move, they have also sent a [pull request](https://github.com/duckdb/duckdb/pull/9635) to DuckDB. The new compression scheme replaces Chimp and Patas. Inside DuckDB, ALP is **x2-4 times faster** than Patas (at decompression) achieving **compression ratios twice as high** (sometimes even much more). -
| Compression | Load | Query | Size | |:-------------|--------:|--------:|-------:| diff --git a/_posts/2024-07-05-community-extensions.md b/_posts/2024-07-05-community-extensions.md index b3b98e1f547..c80af13108c 100644 --- a/_posts/2024-07-05-community-extensions.md +++ b/_posts/2024-07-05-community-extensions.md @@ -96,7 +96,6 @@ For the [maintainer of `h3`](https://github.com/isaacbrodsky/), the publication To show that it’s feasible to publish extensions, we reached out to a few developers of key extensions. At the time of the publication of this blog post, the DuckDB Community Extensions repository already contains the following extensions. -
| Name | Description | |----|------------| diff --git a/docs/api/adbc.md b/docs/api/adbc.md index 15a41c85ab4..e9851df44e0 100644 --- a/docs/api/adbc.md +++ b/docs/api/adbc.md @@ -74,7 +74,6 @@ Functions related to query execution: Functions related to binding, used for bulk insertion or in prepared statements. -
| Function name | Description | Arguments | Example | |:---|:-|:---|:----| diff --git a/docs/api/cli/arguments.md b/docs/api/cli/arguments.md index e40b8d5df64..7b0f951298c 100644 --- a/docs/api/cli/arguments.md +++ b/docs/api/cli/arguments.md @@ -12,7 +12,6 @@ duckdb -help For a list of dot commands available in the CLI shell, see the [Dot Commands page]({% link docs/api/cli/dot_commands.md %}). -
diff --git a/docs/api/cli/dot_commands.md b/docs/api/cli/dot_commands.md index 7eddd31ac46..17761214213 100644 --- a/docs/api/cli/dot_commands.md +++ b/docs/api/cli/dot_commands.md @@ -9,7 +9,6 @@ Dot commands are available in the DuckDB CLI client. To use one of these command ## Dot Commands -
diff --git a/docs/api/cli/editing.md b/docs/api/cli/editing.md index 8a531be2099..d2b7e6f279e 100644 --- a/docs/api/cli/editing.md +++ b/docs/api/cli/editing.md @@ -9,7 +9,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Moving -
| Key | Action | |-----------------|------------------------------------------------------------------------| @@ -30,7 +29,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## History -
| Key | Action | |------------|--------------------------------| @@ -45,7 +43,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Changing Text -
| Key | Action | |-------------------|----------------------------------------------------------| @@ -68,7 +65,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Completing -
| Key | Action | |---------------|--------------------------------------------------------| @@ -78,7 +74,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Miscellaneous -
| Key | Action | |------------|------------------------------------------------------------------------------------| diff --git a/docs/api/cli/output_formats.md b/docs/api/cli/output_formats.md index f85339fb9a3..646b5f56b62 100644 --- a/docs/api/cli/output_formats.md +++ b/docs/api/cli/output_formats.md @@ -8,7 +8,6 @@ redirect_from: The `.mode` [dot command]({% link docs/api/cli/dot_commands.md %}) may be used to change the appearance of the tables returned in the terminal output. In addition to customizing the appearance, these modes have additional benefits. This can be useful for presenting DuckDB output elsewhere by redirecting the terminal [output to a file]({% link docs/api/cli/dot_commands.md %}#output-writing-results-to-a-file). Using the `insert` mode will build a series of SQL statements that can be used to insert the data at a later point. The `markdown` mode is particularly useful for building documentation and the `latex` mode is useful for writing academic papers. -
| Mode | Description | |--------------|---------------------------------------------| diff --git a/docs/api/cli/syntax_highlighting.md b/docs/api/cli/syntax_highlighting.md index a1dac333dba..0241214418b 100644 --- a/docs/api/cli/syntax_highlighting.md +++ b/docs/api/cli/syntax_highlighting.md @@ -14,7 +14,6 @@ Syntax highlighting can also be disabled entirely using the `.highlight off` com Below is a list of components that can be configured. -
| Type | Command | Default color | |-------------------------|-------------|-----------------| @@ -27,7 +26,6 @@ Below is a list of components that can be configured. The components can be configured using either a supported color name (e.g., `.keyword red`), or by directly providing a terminal code to use for rendering (e.g., `.keywordcode \033[31m`). Below is a list of supported color names and their corresponding terminal codes. -
| Color | Terminal code | |---------------|---------------| diff --git a/docs/archive/0.10/api/adbc.md b/docs/archive/0.10/api/adbc.md index d9d16faff6a..711c79e92b0 100644 --- a/docs/archive/0.10/api/adbc.md +++ b/docs/archive/0.10/api/adbc.md @@ -74,7 +74,6 @@ Functions related to query execution: Functions related to binding, used for bulk insertion or in prepared statements. -
| Function name | Description | Arguments | Example | |:---|:-|:---|:----| diff --git a/docs/archive/0.10/api/cli/arguments.md b/docs/archive/0.10/api/cli/arguments.md index 7a44b51dc17..82d63c0bc93 100644 --- a/docs/archive/0.10/api/cli/arguments.md +++ b/docs/archive/0.10/api/cli/arguments.md @@ -7,7 +7,6 @@ The table below summarizes DuckDB's command line options. To list all command line options, use the command `duckdb -help`. Fot a list of dot commands available in the CLI shell, see the [Dot Commands page](dot_commands). -
diff --git a/docs/archive/0.10/api/cli/dot_commands.md b/docs/archive/0.10/api/cli/dot_commands.md index c4b2c546b40..69ab871417c 100644 --- a/docs/archive/0.10/api/cli/dot_commands.md +++ b/docs/archive/0.10/api/cli/dot_commands.md @@ -9,7 +9,6 @@ Dot commands are available in the DuckDB CLI client. To use one of these command ## Dot Commands -
diff --git a/docs/archive/0.10/api/cli/editing.md b/docs/archive/0.10/api/cli/editing.md index aa35847afac..b26572f88e3 100644 --- a/docs/archive/0.10/api/cli/editing.md +++ b/docs/archive/0.10/api/cli/editing.md @@ -9,7 +9,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Moving -
| Key | Action | |-----------------|------------------------------------------------------------------------| @@ -30,7 +29,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## History -
| Key | Action | |------------|--------------------------------| @@ -45,7 +43,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Changing Text -
| Key | Action | |-------------------|----------------------------------------------------------| @@ -68,7 +65,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Completing -
| Key | Action | |---------------|--------------------------------------------------------| @@ -78,7 +74,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Miscellaneous -
| Key | Action | |------------|------------------------------------------------------------------------------------| diff --git a/docs/archive/0.10/api/cli/output_formats.md b/docs/archive/0.10/api/cli/output_formats.md index eaa39dea3aa..9b7e5039799 100644 --- a/docs/archive/0.10/api/cli/output_formats.md +++ b/docs/archive/0.10/api/cli/output_formats.md @@ -8,7 +8,6 @@ title: Output Formats The `.mode` [dot command](dot_commands) may be used to change the appearance of the tables returned in the terminal output. In addition to customizing the appearance, these modes have additional benefits. This can be useful for presenting DuckDB output elsewhere by redirecting the terminal [output to a file](dot_commands#output-writing-results-to-a-file). Using the `insert` mode will build a series of SQL statements that can be used to insert the data at a later point. The `markdown` mode is particularly useful for building documentation and the `latex` mode is useful for writing academic papers. -
| Mode | Description | |--------------|----------------------------------------------| diff --git a/docs/archive/0.10/api/cli/syntax_highlighting.md b/docs/archive/0.10/api/cli/syntax_highlighting.md index d47f32f04fd..a8a82e6d911 100644 --- a/docs/archive/0.10/api/cli/syntax_highlighting.md +++ b/docs/archive/0.10/api/cli/syntax_highlighting.md @@ -14,7 +14,6 @@ Syntax highlighting can also be disabled entirely using the `.highlight off` com Below is a list of components that can be configured. -
| Type | Command | Default Color | |-------------------------|-----------|---------------| @@ -27,7 +26,6 @@ Below is a list of components that can be configured. The components can be configured using either a supported color name (e.g., `.keyword red`), or by directly providing a terminal code to use for rendering (e.g., `.keywordcode \033[31m`). Below is a list of supported color names and their corresponding terminal codes. -
| Color | Terminal Code | |---------------|---------------| diff --git a/docs/archive/0.10/api/python/types.md b/docs/archive/0.10/api/python/types.md index 6489b8be528..0166649298c 100644 --- a/docs/archive/0.10/api/python/types.md +++ b/docs/archive/0.10/api/python/types.md @@ -14,7 +14,6 @@ This means that wherever a DuckDBPyType object is expected, it is also possible The table below shows the mapping of Python Built-in types to DuckDB type. -
| Built-in types | DuckDB type | |:---------------|:------------| @@ -29,7 +28,6 @@ The table below shows the mapping of Python Built-in types to DuckDB type. The table below shows the mapping of Numpy DType to DuckDB type. -
| Type | DuckDB type | |:------------|:------------| @@ -98,7 +96,6 @@ duckdb.typing.DuckDBPyType(Union[int, str, bool, bytearray]) For the built-in types, you can use the constants defined in `duckdb.typing`: -
| DuckDB type | |:---------------| diff --git a/docs/archive/0.10/configuration/pragmas.md b/docs/archive/0.10/configuration/pragmas.md index 9a17a97c781..242f5290c19 100644 --- a/docs/archive/0.10/configuration/pragmas.md +++ b/docs/archive/0.10/configuration/pragmas.md @@ -104,7 +104,6 @@ CALL pragma_storage_info('table_name'); This call returns the following information for the given table: -
| Name | Type | Description | |----------------|-----------|-------------------------------------------------------| diff --git a/docs/archive/0.10/data/csv/auto_detection.md b/docs/archive/0.10/data/csv/auto_detection.md index 6dcb8d0899a..23272da5d55 100644 --- a/docs/archive/0.10/data/csv/auto_detection.md +++ b/docs/archive/0.10/data/csv/auto_detection.md @@ -60,7 +60,6 @@ Dialect detection works by attempting to parse the samples using the set of cons The following dialects are considered for automatic dialect detection. -
@@ -95,7 +94,6 @@ After detecting the dialect, the system will attempt to figure out the types of The type detection works by attempting to convert the values in each column to the candidate types. If the conversion is unsuccessful, the candidate type is removed from the set of candidate types for that column. After all samples have been handled – the remaining candidate type with the highest priority is chosen. The set of considered candidate types in order of priority is given below: -
| Types | |-------------| @@ -131,7 +129,6 @@ If the ambiguities cannot be resolved by looking at the data the system has a li The system considers the following formats for dates (`dateformat`). Higher entries are chosen over lower entries in case of ambiguities (i.e., ISO 8601 is preferred over `MM-DD-YYYY`). -
| dateformat | |------------| @@ -145,7 +142,6 @@ The system considers the following formats for dates (`dateformat`). Higher entr The system considers the following formats for timestamps (`timestampformat`). Higher entries are chosen over lower entries in case of ambiguities. -
| timestampformat | |------------------------| diff --git a/docs/archive/0.10/data/csv/overview.md b/docs/archive/0.10/data/csv/overview.md index a7591790be0..d7bb3adaa6a 100644 --- a/docs/archive/0.10/data/csv/overview.md +++ b/docs/archive/0.10/data/csv/overview.md @@ -130,7 +130,6 @@ The `read_csv` automatically attempts to figure out the correct configuration of SELECT * FROM read_csv('flights.csv'); ``` -
| FlightDate | UniqueCarrier | OriginCityName | DestCityName | |------------|---------------|----------------|-----------------| @@ -147,7 +146,6 @@ CREATE TABLE ontime AS SELECT * FROM read_csv('flights.csv'); DESCRIBE ontime; ``` -
| column_name | column_type | null | key | default | extra | |----------------|-------------|------|------|---------|-------| @@ -189,7 +187,6 @@ COPY ontime FROM 'flights.csv'; SELECT * FROM ontime; ``` -
| flightdate | uniquecarrier | origincityname | destcityname | |------------|---------------|----------------|-----------------| diff --git a/docs/archive/0.10/data/csv/reading_faulty_csv_files.md b/docs/archive/0.10/data/csv/reading_faulty_csv_files.md index 213c672c84f..d52648b2955 100644 --- a/docs/archive/0.10/data/csv/reading_faulty_csv_files.md +++ b/docs/archive/0.10/data/csv/reading_faulty_csv_files.md @@ -112,7 +112,6 @@ FROM read_csv( Outputs: -
| name | age | |-------|-----| @@ -127,7 +126,6 @@ FROM read_csv('faulty.csv', columns = {'name': 'VARCHAR', 'age': 'INTEGER'}); Outputs: -
| name | |--------------| @@ -148,7 +146,6 @@ Note that any of the errors described in our Structural Error section will be st The CSV Reject Scans Table returns the following information: -
| Column name | Description | Type | |:--|:-----|:-| @@ -170,7 +167,6 @@ The CSV Reject Scans Table returns the following information: The CSV Reject Errors Table returns the following information: -
| Column name | Description | Type | |:--|:-----|:-| @@ -187,7 +183,6 @@ The CSV Reject Errors Table returns the following information: ## Parameters -
The parameters listed below are used in the `read_csv` function to configure the CSV Rejects Table. @@ -216,7 +211,6 @@ FROM reject_scans; Outputs: -
| scan_id | file_id | file_path | delimiter | quote | escape | newline_delimiter | skip_rows | has_header | columns | date_format | timestamp_format | user_arguments | |---------|---------|-----------------------------------|-----------|-------|--------|-------------------|-----------|-----------:|--------------------------------------|-------------|------------------|--------------------| @@ -228,7 +222,6 @@ FROM reject_errors; Outputs: -
| scan_id | file_id | line | line_byte_position | byte_position | column_idx | column_name | error_type | csv_line | error_message | |---------|---------|------|--------------------|---------------|------------|-------------|------------|---------------------|------------------------------------------------------------------------------------| diff --git a/docs/archive/0.10/data/json/overview.md b/docs/archive/0.10/data/json/overview.md index 673a5d96ea5..1a9b6a27957 100644 --- a/docs/archive/0.10/data/json/overview.md +++ b/docs/archive/0.10/data/json/overview.md @@ -110,7 +110,6 @@ SELECT * FROM read_json_auto('records.json', format = 'newline_delimited'); ``` -
| key1 | key2 | |----------|----------| @@ -135,7 +134,6 @@ SELECT * FROM read_json_auto('array.json', format = 'array'); ``` -
| key1 | key2 | |----------|----------| @@ -167,7 +165,6 @@ SELECT * FROM read_json_auto('unstructured.json', format = 'unstructured'); ``` -
| key1 | key2 | |----------|----------| @@ -193,7 +190,6 @@ SELECT * FROM read_json_auto('records.json', records = true); ``` -
| key1 | key2 | |----------|----------| @@ -208,7 +204,6 @@ SELECT * FROM read_json_auto('records.json', records = false); ``` -
| json | |------------------------------------| @@ -229,7 +224,6 @@ SELECT * FROM read_json_auto('arrays.json', records = false); ``` -
| json | |-------------| @@ -251,7 +245,6 @@ FROM read_json_auto('todos.json') LIMIT 5; ``` -
| userId | id | title | completed | |--------|----|-----------------------------------------------------------------|-----------| @@ -272,7 +265,6 @@ CREATE TABLE todos AS DESCRIBE todos; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|-----|---------|-------| @@ -302,7 +294,6 @@ COPY todos FROM 'todos.json'; SELECT * FROM todos LIMIT 5; ``` -
| userId | id | title | completed | |--------|----|-----------------------------------------------------------------|-----------| diff --git a/docs/archive/0.10/data/multiple_files/combining_schemas.md b/docs/archive/0.10/data/multiple_files/combining_schemas.md index cec60a473cc..80d9a38b2c4 100644 --- a/docs/archive/0.10/data/multiple_files/combining_schemas.md +++ b/docs/archive/0.10/data/multiple_files/combining_schemas.md @@ -48,7 +48,6 @@ FlightDate|UniqueCarrier|OriginCityName|DestCityName Reading the two files at the same time will produce the following result set: -
| FlightDate | UniqueCarrier | OriginCityName | DestCityName | |------------|---------------|----------------|-----------------| @@ -83,7 +82,6 @@ Reading these when unifying column names **by position** results in an error – SELECT * FROM read_csv(['flights3.csv', 'flights4.csv'], union_by_name = true); ``` -
| FlightDate | OriginCityName | DestCityName | UniqueCarrier | |------------|----------------|-----------------|---------------| diff --git a/docs/archive/0.10/data/multiple_files/overview.md b/docs/archive/0.10/data/multiple_files/overview.md index 2479e5ea21d..3dd13feea3f 100644 --- a/docs/archive/0.10/data/multiple_files/overview.md +++ b/docs/archive/0.10/data/multiple_files/overview.md @@ -84,7 +84,6 @@ SELECT * FROM read_parquet(['file1.parquet', 'file2.parquet', 'file3.parquet']); Any file name input to the `read_parquet` function can either be an exact filename, or use a glob syntax to read multiple files that match a pattern. -
| Wildcard | Description | |------------|-----------------------------------------------------------| @@ -124,7 +123,6 @@ The `filename` argument can be used to add an extra `filename` column to the res SELECT * FROM read_csv(['flights1.csv', 'flights2.csv'], union_by_name = true, filename = true); ``` -
| FlightDate | OriginCityName | DestCityName | UniqueCarrier | filename | |------------|----------------|-----------------|---------------|--------------| @@ -143,7 +141,6 @@ Search the current directory for all files. SELECT * FROM glob('*'); ``` -
| file | |---------------| diff --git a/docs/archive/0.10/data/parquet/metadata.md b/docs/archive/0.10/data/parquet/metadata.md index b87dd2b3841..d7cc5bad53f 100644 --- a/docs/archive/0.10/data/parquet/metadata.md +++ b/docs/archive/0.10/data/parquet/metadata.md @@ -14,7 +14,6 @@ FROM parquet_metadata('test.parquet'); Below is a table of the columns returned by `parquet_metadata`. -
| Field | Type | | ------------------------- | ----------------- | @@ -62,7 +61,6 @@ FROM parquet_schema('test.parquet'); Below is a table of the columns returned by `parquet_schema`. -
| Field | Type | | ----------------- | --------- | @@ -89,7 +87,6 @@ FROM parquet_file_metadata('test.parquet'); Below is a table of the columns returned by `parquet_file_metadata`. -
| Field | Type | | ----------------------------- | --------- | @@ -112,7 +109,6 @@ FROM parquet_kv_metadata('test.parquet'); Below is a table of the columns returned by `parquet_kv_metadata`. -
| Field | Type | | ----------- | --------- | diff --git a/docs/archive/0.10/dev/release_calendar.md b/docs/archive/0.10/dev/release_calendar.md index 14fe842e281..7af1b67f146 100644 --- a/docs/archive/0.10/dev/release_calendar.md +++ b/docs/archive/0.10/dev/release_calendar.md @@ -14,7 +14,6 @@ Patch versions only ship bugfixes, while minor versions also introduce new featu The planned dates of upcoming DuckDB releases are shown below. **Please note that these dates are tentative** and DuckDB maintainers may decide to push back release dates to ensure the stability and quality of releases. -
@@ -34,7 +33,6 @@ The planned dates of upcoming DuckDB releases are shown below. ## Past Releases -
In the following, we list DuckDB's past releases, including their codename (where applicable). Prior to version 0.4.0, all releases, including patch versions, received a codename. diff --git a/docs/archive/0.10/extensions/arrow.md b/docs/archive/0.10/extensions/arrow.md index 2761d01db6b..d2ef9ac3821 100644 --- a/docs/archive/0.10/extensions/arrow.md +++ b/docs/archive/0.10/extensions/arrow.md @@ -18,7 +18,6 @@ LOAD arrow; ## Functions -
| Function | Type | Description | |--|----|-------| diff --git a/docs/archive/0.10/extensions/autocomplete.md b/docs/archive/0.10/extensions/autocomplete.md index 30e436ee13c..53666d8f4e8 100644 --- a/docs/archive/0.10/extensions/autocomplete.md +++ b/docs/archive/0.10/extensions/autocomplete.md @@ -13,7 +13,6 @@ For the behavior of the `autocomplete` extension, see the [documentation of the ## Functions -
| Function | Description | |:----------------------------------|:-----------------------------------------------------| @@ -28,7 +27,6 @@ FROM sql_auto_complete('SEL'); Returns: -
| suggestion | suggestion_start | |-------------|------------------| diff --git a/docs/archive/0.10/extensions/full_text_search.md b/docs/archive/0.10/extensions/full_text_search.md index cc980b4c1cf..510c8536790 100644 --- a/docs/archive/0.10/extensions/full_text_search.md +++ b/docs/archive/0.10/extensions/full_text_search.md @@ -56,7 +56,6 @@ drop_fts_index(input_table) Drops a FTS index for the specified table. -
| Name | Type | Description | |:--|:--|:-----------| diff --git a/docs/archive/0.10/extensions/httpfs/s3api.md b/docs/archive/0.10/extensions/httpfs/s3api.md index f0f01187516..a091e7eb317 100644 --- a/docs/archive/0.10/extensions/httpfs/s3api.md +++ b/docs/archive/0.10/extensions/httpfs/s3api.md @@ -11,7 +11,6 @@ The `httpfs` filesystem is tested with [AWS S3](https://aws.amazon.com/s3/), [Mi The following table shows which parts of the S3 API are required for each `httpfs` feature. -
| Feature | Required S3 API features | |:---|:---| @@ -193,7 +192,6 @@ FROM read_parquet('s3://bucket/*.parquet', filename = true); could for example result in: -
| column_a | column_b | filename | |:---|:---|:---| @@ -241,7 +239,6 @@ s3://my-bucket/partitioned/part_col_a=⟨val⟩/part_col_b=⟨val⟩/data_⟨thr Some additional configuration options exist for the S3 upload, though the default values should suffice for most use cases. -
| Name | Description | |:---|:---| diff --git a/docs/archive/0.10/extensions/httpfs/s3api_legacy_authentication.md b/docs/archive/0.10/extensions/httpfs/s3api_legacy_authentication.md index 6a7116b1e65..ebc70733519 100644 --- a/docs/archive/0.10/extensions/httpfs/s3api_legacy_authentication.md +++ b/docs/archive/0.10/extensions/httpfs/s3api_legacy_authentication.md @@ -73,7 +73,6 @@ Some additional configuration options exist for the S3 upload, though the defaul Additionally, most of the configuration options can be set via environment variables: -
| DuckDB setting | Environment variable | Note | |:-----------------------|:------------------------|:-----------------------------------------| diff --git a/docs/archive/0.10/extensions/json.md b/docs/archive/0.10/extensions/json.md index e84b528cd8e..7bc15603718 100644 --- a/docs/archive/0.10/extensions/json.md +++ b/docs/archive/0.10/extensions/json.md @@ -255,7 +255,6 @@ FROM read_json(['my_file1.json', 'my_file2.json'], columns = {duck: 'INTEGER', goose: 'INTEGER[]', swan: 'DOUBLE'}); ``` -
| duck | goose | swan | |:---|:---|:---| @@ -269,7 +268,6 @@ SELECT goose, duck FROM read_json_auto('*.json.gz'); SELECT goose, duck FROM '*.json.gz'; -- equivalent ``` -
| goose | duck | |:---|:---| @@ -307,7 +305,6 @@ Can be queried exactly the same as a JSON file that contains `'unstructured'` JS Both can be read as the table: -
| duck | goose | |:---|:---| @@ -325,7 +322,6 @@ The `records` parameter specifies whether the JSON contains records that should Results in two columns: -
| duck | goose | |:---|:---| @@ -334,7 +330,6 @@ Results in two columns: You can read the same file with `records` set to `'false'`, to get a single column, which is a `STRUCT` containing the data: -
| json | |:---| @@ -734,7 +729,6 @@ FROM extracted; The following functions are used to create JSON. -
| Function | Description | |:--|:----| @@ -808,7 +802,6 @@ SELECT json_merge_patch('{"duck": 42}', '{"goose": 123}'); There are three JSON aggregate functions. -
| Function | Description | |:---|:----| @@ -859,7 +852,6 @@ SELECT json_group_structure(j) FROM example2; In many cases, it is inefficient to extract values from JSON one-by-one. Instead, we can "extract" all values at once, transforming JSON to the nested types `LIST` and `STRUCT`. -
| Function | Description | |:---|:---| diff --git a/docs/archive/0.10/extensions/mysql.md b/docs/archive/0.10/extensions/mysql.md index c4e52797de2..6fdfa091357 100644 --- a/docs/archive/0.10/extensions/mysql.md +++ b/docs/archive/0.10/extensions/mysql.md @@ -34,7 +34,6 @@ USE mysqldb; The connection string determines the parameters for how to connect to MySQL as a set of `key=value` pairs. Any options not provided are replaced by their default values, as per the table below. Connection information can also be specified with [environment variables](https://dev.mysql.com/doc/refman/8.3/en/environment-variables.html). If no option is provided explicitly, the MySQL extension tries to read it from an environment variable. -
| Setting | Default | Environment variable | |------------|--------------|----------------------| diff --git a/docs/archive/0.10/extensions/official_extensions.md b/docs/archive/0.10/extensions/official_extensions.md index 9fff353efdc..56ec14cd094 100644 --- a/docs/archive/0.10/extensions/official_extensions.md +++ b/docs/archive/0.10/extensions/official_extensions.md @@ -34,7 +34,6 @@ title: Official Extensions Different DuckDB clients ship a different set of extensions. We summarize the main distributions in the table below. -
| Name | CLI (duckdb.org) | CLI (Homebrew) | Python | R | Java | Node.js | |------|------|------|---|---|---|---|---| diff --git a/docs/archive/0.10/extensions/overview.md b/docs/archive/0.10/extensions/overview.md index c480632fd2e..47a26962553 100644 --- a/docs/archive/0.10/extensions/overview.md +++ b/docs/archive/0.10/extensions/overview.md @@ -23,7 +23,6 @@ SELECT extension_name, installed, description FROM duckdb_extensions(); ``` -
| extension_name | installed | description | |-------------------|-----------|--------------------------------------------------------------| diff --git a/docs/archive/0.10/extensions/spatial.md b/docs/archive/0.10/extensions/spatial.md index 50b2a913b38..95c186cf8ed 100644 --- a/docs/archive/0.10/extensions/spatial.md +++ b/docs/archive/0.10/extensions/spatial.md @@ -138,7 +138,6 @@ Compute relationships and spatial predicates between geometries. ## Spatial Aggregate Functions -
| Aggregate functions | Implemented with | |-------------------------------------------|------------------| diff --git a/docs/archive/0.10/extensions/tpch.md b/docs/archive/0.10/extensions/tpch.md index a579818a29e..5adf4812bb5 100644 --- a/docs/archive/0.10/extensions/tpch.md +++ b/docs/archive/0.10/extensions/tpch.md @@ -80,7 +80,6 @@ This function returns a table with columns `query_nr`, `scale_factor`, and `answ The data generator function `dbgen` has the following parameters: -
| Name | Type | Description | |--|--|------------| diff --git a/docs/archive/0.10/extensions/versioning_of_extensions.md b/docs/archive/0.10/extensions/versioning_of_extensions.md index 0324329c517..2575776b2fd 100644 --- a/docs/archive/0.10/extensions/versioning_of_extensions.md +++ b/docs/archive/0.10/extensions/versioning_of_extensions.md @@ -13,7 +13,6 @@ For example, the parquet extension is built into DuckDB version `v0.10.3` (which ```sql SELECT extension_name, extension_version, install_mode FROM duckdb_extensions() WHERE extension_name='parquet'; ``` -
| extension_name | extension_version | install_mode | |:------------------|:------------------|:---------------------| @@ -30,7 +29,6 @@ FROM duckdb_extensions() WHERE extension_name = 'azure'; ``` -
| extension_name | extension_version | install_mode | |:---------------|:------------------|:---------------| @@ -47,7 +45,6 @@ give the user information on which extensions were updated to/from which version UPDATE EXTENSIONS; ``` -
| extension_name | repository | update_result | previous_version | current_version | |:---------------|:-------------|:----------------------|:-----------------|:----------------| @@ -65,7 +62,6 @@ The update statement can also be provided with a list of specific extensions to UPDATE EXTENSIONS (httpfs, azure); ``` -
| extension_name | repository | update_result | previous_version | current_version | |:---------------|:-------------|:----------------------|:-----------------|:----------------| diff --git a/docs/archive/0.10/extensions/working_with_extensions.md b/docs/archive/0.10/extensions/working_with_extensions.md index 4ba784d9fbf..dc0b33faa3a 100644 --- a/docs/archive/0.10/extensions/working_with_extensions.md +++ b/docs/archive/0.10/extensions/working_with_extensions.md @@ -10,7 +10,6 @@ For platforms where packages for certain extensions are not available, users can All official extensions are distributed for the following platforms. -
| Platform name | Description | |--------------------|------------------------------------------| @@ -89,7 +88,6 @@ SET custom_extension_repository = 'http://nightly-extensions.duckdb.org'; While any url or local path can be used as a repository, currently DuckDB contains the following predefined repositories: -
| alias | Url | Description | |:--------------------|:---------------------------------------|:---------------------------------------------------------------------------------------| diff --git a/docs/archive/0.10/guides/database_integration/mysql.md b/docs/archive/0.10/guides/database_integration/mysql.md index bd661512ae4..ba4f8fad874 100644 --- a/docs/archive/0.10/guides/database_integration/mysql.md +++ b/docs/archive/0.10/guides/database_integration/mysql.md @@ -32,7 +32,6 @@ USE mysql_db; The string used by `ATTACH` is a PostgreSQL-style connection string (_not_ a MySQL connection string!). It is a list of connection arguments provided in `{key}={value}` format. Below is a list of valid arguments. Any options not provided are replaced by their default values. -
| Setting | Default | |------------|--------------| diff --git a/docs/archive/0.10/guides/database_integration/postgres.md b/docs/archive/0.10/guides/database_integration/postgres.md index c78ac9370af..a832ecfe35e 100644 --- a/docs/archive/0.10/guides/database_integration/postgres.md +++ b/docs/archive/0.10/guides/database_integration/postgres.md @@ -32,7 +32,6 @@ SELECT * FROM postgres_scan('host=localhost port=5432 dbname=mydb', 'public', 'm The first parameter to the `postgres_scan` function is the [PostgreSQL connection string](https://www.postgresql.org/docs/current/libpq-connect.html#LIBPQ-CONNSTRING), a list of connection arguments provided in `{key}={value}` format. Below is a list of valid arguments. -
| Name | Description | Default | diff --git a/docs/archive/0.10/guides/meta/describe.md b/docs/archive/0.10/guides/meta/describe.md index 203baf32690..58ef0286c0a 100644 --- a/docs/archive/0.10/guides/meta/describe.md +++ b/docs/archive/0.10/guides/meta/describe.md @@ -13,7 +13,6 @@ DESCRIBE tbl; SHOW tbl; -- equivalent to DESCRIBE tbl; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|------|---------|-------| @@ -28,7 +27,6 @@ In order to view the schema of the result of a query, prepend `DESCRIBE` to a qu DESCRIBE SELECT * FROM tbl; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|------|---------|-------| diff --git a/docs/archive/0.10/guides/meta/list_tables.md b/docs/archive/0.10/guides/meta/list_tables.md index 71411bf77db..3180f975076 100644 --- a/docs/archive/0.10/guides/meta/list_tables.md +++ b/docs/archive/0.10/guides/meta/list_tables.md @@ -10,7 +10,6 @@ CREATE TABLE tbl (i INTEGER); SHOW TABLES; ``` -
| name | |------| @@ -25,7 +24,6 @@ CREATE TABLE s1.tbl (v VARCHAR); SHOW ALL TABLES; ``` -
| database | schema | table_name | column_names | column_types | temporary | |----------|--------|------------|--------------|--------------|-----------| diff --git a/docs/archive/0.10/guides/network_cloud_storage/s3_express_one.md b/docs/archive/0.10/guides/network_cloud_storage/s3_express_one.md index e68d46f8eb1..2698f53f65b 100644 --- a/docs/archive/0.10/guides/network_cloud_storage/s3_express_one.md +++ b/docs/archive/0.10/guides/network_cloud_storage/s3_express_one.md @@ -57,7 +57,6 @@ FROM 's3://express-bucket-name--use1-az5--x-s3/my-file.parquet'; We ran two experiments on a `c7gd.12xlarge` instance using the [LDBC SF300 Comments `creationDate` Parquet file](https://blobs.duckdb.org/data/ldbc-sf300-comments-creationDate.parquet) file (also used in the [microbenchmarks of the performance guide](../performance/benchmarks#data-sets)). -
| Experiment | File size | Runtime | |:-----|--:|--:| diff --git a/docs/archive/0.10/guides/performance/file_formats.md b/docs/archive/0.10/guides/performance/file_formats.md index a3940e0620e..21ff4764415 100644 --- a/docs/archive/0.10/guides/performance/file_formats.md +++ b/docs/archive/0.10/guides/performance/file_formats.md @@ -40,7 +40,6 @@ DuckDB works best on Parquet files with row groups of 100K-1M rows each. The rea We run a simple aggregation query over Parquet files using different row group sizes, selected between 960 and 1,966,080. The results are as follows. -
| Row group size | Execution time | |---------------:|---------------:| @@ -77,7 +76,6 @@ For tips on reading and writing Parquet files, see the [Parquet Tips page](../.. CSV files are often distributed in compressed format such as GZIP archives (`.csv.gz`). DuckDB can decompress these files on the fly. In fact, this is typically faster than decompressing the files first and loading them due to reduced IO. -
| Schema | Load Time | |---|--:| diff --git a/docs/archive/0.10/guides/performance/indexing.md b/docs/archive/0.10/guides/performance/indexing.md index 9a13663489a..1c32c100568 100644 --- a/docs/archive/0.10/guides/performance/indexing.md +++ b/docs/archive/0.10/guides/performance/indexing.md @@ -17,7 +17,6 @@ The more ordered the data within a column, the more useful the zonemap indexes w For an example, let’s repeat the [microbenchmark for timestamps](schema#microbenchmark-using-timestamps) with a timestamp column that sorted using an ascending order vs. an unordered one. -
| Column type | Ordered | Storage size | Query time | |---|---|---|---| diff --git a/docs/archive/0.10/guides/performance/schema.md b/docs/archive/0.10/guides/performance/schema.md index 0bee4ed200c..0c0376a8ddc 100644 --- a/docs/archive/0.10/guides/performance/schema.md +++ b/docs/archive/0.10/guides/performance/schema.md @@ -31,7 +31,6 @@ SELECT avg(CAST(creationDate[9:10] AS INTEGER)) FROM Comment; The results of the microbenchmark are as follows: -
| Column type | Storage size | Query time | | ----------- | -----------: | ---------: | @@ -55,7 +54,6 @@ In the second experiment, we define all columns with the `VARCHAR` type. While the results of the queries are the same for all both experiments, their runtime vary significantly. The results below show that joining on `BIGINT` columns is approx. 1.8× faster than performing the same join on `VARCHAR`-typed columns encoding the same value. -
| Join column payload type | Join column schema type | Example value | Query time | | ------------------------ | ----------------------- | ---------------------------------------- | ---------: | @@ -74,7 +72,6 @@ DuckDB allows defining [constraints](../../sql/constraints) such as `UNIQUE`, `P We illustrate the effect of using primary keys with the [LDBC Comment table at scale factor 300](https://blobs.duckdb.org/data/ldbc-sf300-comments.tar.zst). This table has approx. 554 million entries. We first create the schema without a primary key, then load the data. In the second experiment, we create the schema with a primary key, then load the data. In both cases, we take the data from `.csv.gz` files, and measure the time required to perform the loading. -
| Operation | Execution time | | ------------------------ | -------------: | diff --git a/docs/archive/0.10/guides/sql_features/asof_join.md b/docs/archive/0.10/guides/sql_features/asof_join.md index 14405168c85..5859c94f008 100644 --- a/docs/archive/0.10/guides/sql_features/asof_join.md +++ b/docs/archive/0.10/guides/sql_features/asof_join.md @@ -24,7 +24,6 @@ which can be cumbersome and slow to implement in standard SQL. Let's start with a concrete example. Suppose we have a table of stock [`prices`](/data/prices.csv) with timestamps: -
| ticker | when | price | | :----- | :--- | ----: | @@ -40,7 +39,6 @@ Suppose we have a table of stock [`prices`](/data/prices.csv) with timestamps: We have another table containing portfolio [`holdings`](/data/holdings.csv) at various points in time: -
| ticker | when | shares | | :----- | :--- | -----: | @@ -76,7 +74,6 @@ ASOF JOIN prices p This attaches the value of the holding at that time to each row: -
| ticker | when | value | | :----- | :--- | ----: | @@ -107,7 +104,6 @@ ORDER BY ALL; As you might expect, this will produce `NULL` prices and values instead of dropping left side rows when there is no ticker or the time is before the prices begin. -
| ticker | when | value | | :----- | :--- | ----: | diff --git a/docs/archive/0.10/internals/storage.md b/docs/archive/0.10/internals/storage.md index f7ed4b6ee26..c50e1a3fa63 100644 --- a/docs/archive/0.10/internals/storage.md +++ b/docs/archive/0.10/internals/storage.md @@ -65,7 +65,6 @@ with open('test/sql/storage_version/storage_version.db', 'rb') as fh: For changes in each given release, check out the [change log](https://github.com/duckdb/duckdb/releases) on GitHub. To see the commits that changed each storage version, see the [commit log](https://github.com/duckdb/duckdb/commits/main/src/storage/storage_info.cpp). -
| Storage version | DuckDB version(s) | |----------------:|---------------------------------| diff --git a/docs/archive/0.10/sql/data_types/bitstring.md b/docs/archive/0.10/sql/data_types/bitstring.md index 43436322e7d..1a0983134ba 100644 --- a/docs/archive/0.10/sql/data_types/bitstring.md +++ b/docs/archive/0.10/sql/data_types/bitstring.md @@ -4,7 +4,6 @@ layout: docu title: Bitstring Type --- -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/0.10/sql/data_types/blob.md b/docs/archive/0.10/sql/data_types/blob.md index cbcda021590..7c983aca1d7 100644 --- a/docs/archive/0.10/sql/data_types/blob.md +++ b/docs/archive/0.10/sql/data_types/blob.md @@ -5,7 +5,6 @@ layout: docu title: Blob Type --- -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/0.10/sql/data_types/boolean.md b/docs/archive/0.10/sql/data_types/boolean.md index 8f42ff3ea4f..122a390068a 100644 --- a/docs/archive/0.10/sql/data_types/boolean.md +++ b/docs/archive/0.10/sql/data_types/boolean.md @@ -4,7 +4,6 @@ layout: docu title: Boolean Type --- -
| Name | Aliases | Description | |:---|:---|:---| @@ -45,7 +44,6 @@ The `AND`/`OR` conjunctions can be used to combine boolean values. Below is the truth table for the `AND` conjunction (i.e., `x AND y`). -
| X | X AND true | X AND false | X AND NULL | |-------|-------|-------|-------| @@ -55,7 +53,6 @@ Below is the truth table for the `AND` conjunction (i.e., `x AND y`). Below is the truth table for the `OR` conjunction (i.e., `x OR y`). -
| X | X OR true | X OR false | X OR NULL | |-------|------|-------|------| diff --git a/docs/archive/0.10/sql/data_types/date.md b/docs/archive/0.10/sql/data_types/date.md index 94bf44d0725..7057d2ea277 100644 --- a/docs/archive/0.10/sql/data_types/date.md +++ b/docs/archive/0.10/sql/data_types/date.md @@ -4,7 +4,6 @@ layout: docu title: Date Types --- -
| Name | Aliases | Description | |:-------|:--------|:--------------------------------| @@ -20,7 +19,6 @@ SELECT DATE '1992-09-20'; There are also three special date values that can be used on input: -
| Input string | Description | |:-------------|:----------------------------------| diff --git a/docs/archive/0.10/sql/data_types/enum.md b/docs/archive/0.10/sql/data_types/enum.md index d7fc005fb61..8e7c3598e33 100644 --- a/docs/archive/0.10/sql/data_types/enum.md +++ b/docs/archive/0.10/sql/data_types/enum.md @@ -5,7 +5,6 @@ layout: docu title: Enum Data Type --- -
| Name | Description | |:--|:-----| @@ -76,7 +75,6 @@ Show the available values in the `birds` enum using the `enum_range` function: SELECT enum_range(NULL::birds) AS my_enum_range; ``` -
| my_enum_range | |-----------------| diff --git a/docs/archive/0.10/sql/data_types/interval.md b/docs/archive/0.10/sql/data_types/interval.md index 50b95b7c4b9..208d9c8ba7d 100644 --- a/docs/archive/0.10/sql/data_types/interval.md +++ b/docs/archive/0.10/sql/data_types/interval.md @@ -7,7 +7,6 @@ title: Interval Type Intervals represent a period of time. This period can be measured in a specific unit or combination of units, for example years, days, or seconds. Intervals are generally used to *modify* timestamps or dates by either adding or subtracting them. -
| Name | Description | |:---|:---| diff --git a/docs/archive/0.10/sql/data_types/list.md b/docs/archive/0.10/sql/data_types/list.md index 7347b10c871..44015d2b629 100644 --- a/docs/archive/0.10/sql/data_types/list.md +++ b/docs/archive/0.10/sql/data_types/list.md @@ -53,7 +53,6 @@ Retrieving one or more values from a list can be accomplished using brackets and > This is only needed in our basic examples here, not when working with a list column. > For example, this can't be parsed: `SELECT ['a', 'b', 'c'][1]`. -
| Example | Result | |:-------------------------------------------|:-------------| diff --git a/docs/archive/0.10/sql/data_types/numeric.md b/docs/archive/0.10/sql/data_types/numeric.md index d5e66752b4a..5c9c3ecd69d 100644 --- a/docs/archive/0.10/sql/data_types/numeric.md +++ b/docs/archive/0.10/sql/data_types/numeric.md @@ -30,7 +30,6 @@ The data type `DECIMAL(WIDTH, SCALE)` (also available under the alias `NUMERIC(W Internally, decimals are represented as integers depending on their specified width. -
| Width | Internal | Size (bytes) | |:---|:---|---:| @@ -45,7 +44,6 @@ Performance can be impacted by using too large decimals when not required. In pa The data types `REAL` and `DOUBLE` precision are variable-precision numeric types. In practice, these types are usually implementations of IEEE Standard 754 for Binary Floating-Point Arithmetic (single and double precision, respectively), to the extent that the underlying processor, operating system, and compiler support it. -
| Name | Aliases | Description | |:--|:--|:--------| diff --git a/docs/archive/0.10/sql/data_types/struct.md b/docs/archive/0.10/sql/data_types/struct.md index e4ac9bfaf8e..2301f5fd0be 100644 --- a/docs/archive/0.10/sql/data_types/struct.md +++ b/docs/archive/0.10/sql/data_types/struct.md @@ -116,7 +116,6 @@ SELECT a.* FROM (SELECT {'x': 1, 'y': 2, 'z': 3} AS a); ``` -
| x | y | z | |:---|:---|:---| diff --git a/docs/archive/0.10/sql/data_types/text.md b/docs/archive/0.10/sql/data_types/text.md index a873d54497e..7d35d8afd80 100644 --- a/docs/archive/0.10/sql/data_types/text.md +++ b/docs/archive/0.10/sql/data_types/text.md @@ -7,7 +7,6 @@ title: Text Types In DuckDB, strings can be stored in the `VARCHAR` field. The field allows storage of Unicode characters. Internally, the data is encoded as UTF-8. -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/0.10/sql/data_types/time.md b/docs/archive/0.10/sql/data_types/time.md index bb6552fab29..4a35b258759 100644 --- a/docs/archive/0.10/sql/data_types/time.md +++ b/docs/archive/0.10/sql/data_types/time.md @@ -6,7 +6,6 @@ title: Time Types The `TIME` and `TIMETZ` types specify the hour, minute, second, microsecond of a day. -
| Name | Aliases | Description | | :------- | :----------------------- | :------------------------------ | diff --git a/docs/archive/0.10/sql/data_types/timestamp.md b/docs/archive/0.10/sql/data_types/timestamp.md index 788ee20db9e..1f511823186 100644 --- a/docs/archive/0.10/sql/data_types/timestamp.md +++ b/docs/archive/0.10/sql/data_types/timestamp.md @@ -82,7 +82,6 @@ SELECT TIMESTAMP WITH TIME ZONE '1992-09-20 11:30:00.123456789'; There are also three special date values that can be used on input: -
| Input string | Valid types | Description | |:-------------|:---------------------------|:-----------------------------------------------| @@ -97,7 +96,6 @@ but `epoch` is simply a notational shorthand that will be converted to the time SELECT '-infinity'::TIMESTAMP, 'epoch'::TIMESTAMP, 'infinity'::TIMESTAMP; ``` -
| Negative | Epoch | Positive | |:----------|:--------------------|:---------| diff --git a/docs/archive/0.10/sql/data_types/timezones.md b/docs/archive/0.10/sql/data_types/timezones.md index e81a9a738c6..3c0b3409572 100644 --- a/docs/archive/0.10/sql/data_types/timezones.md +++ b/docs/archive/0.10/sql/data_types/timezones.md @@ -12,7 +12,6 @@ FROM pg_timezone_names() ORDER BY name; ``` -
| name | abbrev | |----------------------------------|----------------------------------| diff --git a/docs/archive/0.10/sql/data_types/union.md b/docs/archive/0.10/sql/data_types/union.md index 5f531729b88..75487c461bb 100644 --- a/docs/archive/0.10/sql/data_types/union.md +++ b/docs/archive/0.10/sql/data_types/union.md @@ -95,7 +95,6 @@ So how do we disambiguate if we want to create a `UNION` with multiple members o `UNION` types can be cast between each other if the source type is a "subset" of the target type. In other words, all the tags in the source `UNION` must be present in the target `UNION`, and all the types of the matching tags must be implicitly castable between source and target. In essence, this means that `UNION` types are covariant with respect to their members. -
| Ok | Source | Target | Comments | |----|------------------------|------------------------|----------------------------------------| diff --git a/docs/archive/0.10/sql/duckdb_table_functions.md b/docs/archive/0.10/sql/duckdb_table_functions.md index ea69afe97e9..1d7c4b70b04 100644 --- a/docs/archive/0.10/sql/duckdb_table_functions.md +++ b/docs/archive/0.10/sql/duckdb_table_functions.md @@ -166,7 +166,6 @@ Note that `duckdb_indexes` only provides metadata about secondary indexes – i. The `duckdb_keywords()` function provides metadata about DuckDB's keywords and reserved words. -
| Column | Description | Type | |:-|:---|:-| @@ -188,7 +187,6 @@ The `duckdb_memory()` function provides metadata about DuckDB's buffer manager. The `duckdb_optimizers()` function provides metadata about the optimization rules (e.g., `expression_rewriter`, `filter_pushdown`) available in the DuckDB instance. These can be selectively turned off using [`PRAGMA disabled_optimizers`](../configuration/pragmas#selectively-disabling-optimizers). -
| Column | Description | Type | |:-|:---|:-| @@ -254,7 +252,6 @@ Attributes like `temporary`, `start_value` etc. correspond to the various option The `duckdb_settings()` function provides metadata about the settings available in the DuckDB instance. -
| Column | Description | Type | |:-|:---|:-| @@ -292,7 +289,6 @@ The [`information_schema.tables`](information_schema#tables-and-views) system vi The `duckdb_temporary_files()` function provides metadata about the temporary files DuckDB has written to disk, to offload data from memory. This function mostly exists for debugging and testing purposes. -
| Column | Description | Type | |:-|:---|:-| diff --git a/docs/archive/0.10/sql/expressions/comparison_operators.md b/docs/archive/0.10/sql/expressions/comparison_operators.md index 593a8b1735d..f376fd02dc7 100644 --- a/docs/archive/0.10/sql/expressions/comparison_operators.md +++ b/docs/archive/0.10/sql/expressions/comparison_operators.md @@ -11,7 +11,6 @@ title: Comparisons The table below shows the standard comparison operators. Whenever either of the input arguments is `NULL`, the output of the comparison is `NULL`. -
| Operator | Description | Example | Result | |:---|:---|:---|:---| @@ -25,7 +24,6 @@ Whenever either of the input arguments is `NULL`, the output of the comparison i The table below shows the standard distinction operators. These operators treat `NULL` values as equal. -
| Operator | Description | Example | Result | |:---|:---|:---|:-| @@ -40,7 +38,6 @@ Besides the standard comparison operators there are also the `BETWEEN` and `IS ( Note that `BETWEEN` and `NOT BETWEEN` are only equivalent to the examples below in the cases where both `a`, `x` and `y` are of the same type, as `BETWEEN` will cast all of its inputs to the same type. -
| Predicate | Description | |:---|:---| diff --git a/docs/archive/0.10/sql/expressions/logical_operators.md b/docs/archive/0.10/sql/expressions/logical_operators.md index 57a6a4f1f88..5249033ac8b 100644 --- a/docs/archive/0.10/sql/expressions/logical_operators.md +++ b/docs/archive/0.10/sql/expressions/logical_operators.md @@ -10,7 +10,6 @@ The following logical operators are available: `AND`, `OR` and `NOT`. SQL uses a ### Binary Operators: `AND` and `OR` -
| `a` | `b` | `a AND b` | `a OR b` | |:---|:---|:---|:---| @@ -23,7 +22,6 @@ The following logical operators are available: `AND`, `OR` and `NOT`. SQL uses a ### Unary Operator: `NOT` -
| `a` | `NOT a` | |:---|:---| diff --git a/docs/archive/0.10/sql/expressions/star.md b/docs/archive/0.10/sql/expressions/star.md index 7b7d7fddf55..a1e837bae45 100644 --- a/docs/archive/0.10/sql/expressions/star.md +++ b/docs/archive/0.10/sql/expressions/star.md @@ -91,7 +91,6 @@ INSERT INTO numbers VALUES (1, 10), (2, 20), (3, NULL); SELECT min(COLUMNS(*)), count(COLUMNS(*)) FROM numbers; ``` -
| id | number | id | number | |---:|-------:|---:|-------:| @@ -106,7 +105,6 @@ SELECT FROM numbers; ``` -
| id | min(number := (number + id)) | id | |---:|-----------------------------:|---:| @@ -118,7 +116,6 @@ FROM numbers; SELECT COLUMNS(*) + COLUMNS(*) FROM numbers; ``` -
| id | number | |---:|-------:| @@ -140,7 +137,6 @@ FROM ( WHERE COLUMNS(*) > 1; -- equivalent to: x > 1 AND y > 1 AND z > 1 ``` -
| x | y | z | |--:|--:|--:| @@ -154,7 +150,6 @@ WHERE COLUMNS(*) > 1; -- equivalent to: x > 1 AND y > 1 AND z > 1 SELECT COLUMNS('(id|numbers?)') FROM numbers; ``` -
| id | number | |---:|-------:| @@ -168,7 +163,6 @@ The matches of capture groups can be used to rename columns selected by a regula SELECT COLUMNS('(\w{2}).*') AS '\1' FROM numbers; ``` -
| id | nu | |---:|-----:| @@ -186,7 +180,6 @@ The capture groups are one-indexed; `\0` is the original column name. SELECT COLUMNS(c -> c LIKE '%num%') FROM numbers; ``` -
| number | |--------| @@ -206,7 +199,6 @@ For example: SELECT st.* FROM (SELECT {'x': 1, 'y': 2, 'z': 3} AS st); ``` -
| x | y | z | |--:|--:|--:| diff --git a/docs/archive/0.10/sql/expressions/subqueries.md b/docs/archive/0.10/sql/expressions/subqueries.md index e2ccd30ac9e..ec04177a2ed 100644 --- a/docs/archive/0.10/sql/expressions/subqueries.md +++ b/docs/archive/0.10/sql/expressions/subqueries.md @@ -16,7 +16,6 @@ Consider the following table: ### Grades -
| grade | course | |---:|:---| diff --git a/docs/archive/0.10/sql/functions/bitstring.md b/docs/archive/0.10/sql/functions/bitstring.md index 8025d96c411..349a7b28766 100644 --- a/docs/archive/0.10/sql/functions/bitstring.md +++ b/docs/archive/0.10/sql/functions/bitstring.md @@ -12,7 +12,6 @@ Bitstrings must be of equal length when performing the bitwise operands AND, OR The table below shows the available mathematical operators for `BIT` type. -
diff --git a/docs/archive/0.10/sql/functions/char.md b/docs/archive/0.10/sql/functions/char.md index 625e28ec622..45baab9dbee 100644 --- a/docs/archive/0.10/sql/functions/char.md +++ b/docs/archive/0.10/sql/functions/char.md @@ -842,7 +842,6 @@ SELECT format('I''d rather be {1} than {0}.', 'right', 'happy'); -- I'd rather b #### Format Specifiers -
| Specifier | Description | Example | |:-|:------|:---| @@ -953,7 +952,6 @@ I'd rather be happy than right. #### Format Specifiers -
| Specifier | Description | Example | |:-|:------|:---| diff --git a/docs/archive/0.10/sql/functions/dateformat.md b/docs/archive/0.10/sql/functions/dateformat.md index cb70a425eda..2cb7f741249 100644 --- a/docs/archive/0.10/sql/functions/dateformat.md +++ b/docs/archive/0.10/sql/functions/dateformat.md @@ -66,7 +66,6 @@ FROM read_csv('test.csv', dateformat = '%m/%d/%Y'); Below is a full list of all available format specifiers. -
| Specifier | Description | Example | |:-|:------|:---| diff --git a/docs/archive/0.10/sql/functions/numeric.md b/docs/archive/0.10/sql/functions/numeric.md index b64b0a6870d..e378f65ba68 100644 --- a/docs/archive/0.10/sql/functions/numeric.md +++ b/docs/archive/0.10/sql/functions/numeric.md @@ -9,7 +9,6 @@ title: Numeric Functions The table below shows the available mathematical operators for numeric types. -
diff --git a/docs/archive/0.10/sql/functions/pattern_matching.md b/docs/archive/0.10/sql/functions/pattern_matching.md index 42650530d10..5bfdd5dd555 100644 --- a/docs/archive/0.10/sql/functions/pattern_matching.md +++ b/docs/archive/0.10/sql/functions/pattern_matching.md @@ -64,7 +64,6 @@ SELECT 'A%c' ILIKE 'a$%c' ESCAPE '$'; -- true There are also alternative characters that can be used as keywords in place of `LIKE` expressions. These enhance PostgreSQL compatibility. -
| LIKE-style | PostgreSQL-style | |:---|:---| @@ -93,7 +92,6 @@ SELECT 'abc' NOT SIMILAR TO 'abc'; -- false There are also alternative characters that can be used as keywords in place of `SIMILAR TO` expressions. These follow POSIX syntax. -
| `SIMILAR TO`-style | POSIX-style | |:---|:---| @@ -138,7 +136,6 @@ SELECT NOT 'best.txt' GLOB '*.txt'; -- false Three tildes (`~~~`) may also be used in place of the `GLOB` keyword. -
| GLOB-style | Symbolic-style | |:---|:---| @@ -155,7 +152,6 @@ Search the current directory for all files: SELECT * FROM glob('*'); ``` -
| file | |---------------| diff --git a/docs/archive/0.10/sql/functions/regular_expressions.md b/docs/archive/0.10/sql/functions/regular_expressions.md index ff3a53825cd..fa13ee41f7a 100644 --- a/docs/archive/0.10/sql/functions/regular_expressions.md +++ b/docs/archive/0.10/sql/functions/regular_expressions.md @@ -110,7 +110,6 @@ SELECT regexp_matches('abc', 'A', 'i'); -- true The regex functions support the following `options`. -
| Option | Description | |:---|:---| @@ -135,7 +134,6 @@ SELECT regexp_matches(E'hello\nworld', 'hello.world', 's'); -- true The `regexp_matches` operator will be optimized to the `LIKE` operator when possible. To achieve best performance, the `'c'` option (case-sensitive matching) should be passed if applicable. Note that by default the [`RE2` library](#regular-expression-syntax) doesn't match the `.` character to newline. -
| Original | Optimized equivalent | |:---|:---| diff --git a/docs/archive/0.10/sql/functions/time.md b/docs/archive/0.10/sql/functions/time.md index fc30c4f74e4..078d61bfd17 100644 --- a/docs/archive/0.10/sql/functions/time.md +++ b/docs/archive/0.10/sql/functions/time.md @@ -9,7 +9,6 @@ This section describes functions and operators for examining and manipulating `T The table below shows the available mathematical operators for `TIME` types. -
| Operator | Description | Example | Result | |:-|:---|:----|:--| diff --git a/docs/archive/0.10/sql/functions/utility.md b/docs/archive/0.10/sql/functions/utility.md index cfc1412f45e..cd9fc467e4b 100644 --- a/docs/archive/0.10/sql/functions/utility.md +++ b/docs/archive/0.10/sql/functions/utility.md @@ -267,7 +267,6 @@ The functions below are difficult to categorize into specific function types and A table function is used in place of a table in a `FROM` clause. -
| Name | Description | |:--|:-------| diff --git a/docs/archive/0.10/sql/query_syntax/filter.md b/docs/archive/0.10/sql/query_syntax/filter.md index 8b49aba85a6..4330b4f984b 100644 --- a/docs/archive/0.10/sql/query_syntax/filter.md +++ b/docs/archive/0.10/sql/query_syntax/filter.md @@ -26,7 +26,6 @@ SELECT FROM generate_series(1, 10) tbl(i); ``` -
| total_rows | lte_five | odds | |:---|:---|:---| @@ -42,7 +41,6 @@ SELECT FROM generate_series(1, 10) tbl(i); ``` -
| lte_five_sum | odds_median | odds_lte_five_median | |:---|:---|:---| @@ -94,7 +92,6 @@ SELECT FROM stacked_data; ``` -
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| @@ -114,7 +111,6 @@ SELECT FROM stacked_data; ``` -
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| @@ -132,7 +128,6 @@ SELECT FROM stacked_data; ``` -
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| diff --git a/docs/archive/0.10/sql/query_syntax/orderby.md b/docs/archive/0.10/sql/query_syntax/orderby.md index bfc328a2b9a..d14cb333b96 100644 --- a/docs/archive/0.10/sql/query_syntax/orderby.md +++ b/docs/archive/0.10/sql/query_syntax/orderby.md @@ -104,7 +104,6 @@ FROM addresses ORDER BY ALL; ``` -
| address | city | zip | |------------------------|-----------|------------| @@ -121,7 +120,6 @@ FROM addresses ORDER BY ALL DESC; ``` -
| address | city | zip | |------------------------|-----------|------------| diff --git a/docs/archive/0.10/sql/query_syntax/qualify.md b/docs/archive/0.10/sql/query_syntax/qualify.md index ec27437bdef..2045efda211 100644 --- a/docs/archive/0.10/sql/query_syntax/qualify.md +++ b/docs/archive/0.10/sql/query_syntax/qualify.md @@ -88,7 +88,6 @@ WHERE function_rank < 3; ``` -
| schema_name | function_name | function_rank | |:---|:---|:---| diff --git a/docs/archive/0.10/sql/statements/create_index.md b/docs/archive/0.10/sql/statements/create_index.md index a1832fa0eaa..7405be3041f 100644 --- a/docs/archive/0.10/sql/statements/create_index.md +++ b/docs/archive/0.10/sql/statements/create_index.md @@ -38,7 +38,6 @@ CREATE INDEX i_index ON integers ((j + k)); ### Parameters -
| Name | Description | |:-|:-----| @@ -68,7 +67,6 @@ DROP INDEX title_idx; ### Parameters -
| Name | Description | |:---|:---| diff --git a/docs/archive/0.10/sql/statements/insert.md b/docs/archive/0.10/sql/statements/insert.md index a848ee91285..f494d78204a 100644 --- a/docs/archive/0.10/sql/statements/insert.md +++ b/docs/archive/0.10/sql/statements/insert.md @@ -388,7 +388,6 @@ INSERT INTO t1 RETURNING *; ``` -
| i | |---:| @@ -403,7 +402,6 @@ INSERT INTO t2 RETURNING *, i * j AS i_times_j; ``` -
| i | j | i_times_j | |--:|--:|----------:| @@ -421,7 +419,6 @@ INSERT INTO t3 RETURNING *; ``` -
| i | j | |--:|---:| diff --git a/docs/archive/0.10/sql/statements/pivot.md b/docs/archive/0.10/sql/statements/pivot.md index d67df8f36d7..afe69ce5582 100644 --- a/docs/archive/0.10/sql/statements/pivot.md +++ b/docs/archive/0.10/sql/statements/pivot.md @@ -50,7 +50,6 @@ INSERT INTO Cities VALUES ('US', 'New York City', 2020, 8772); FROM Cities; ``` -
| Country | Name | Year | Population | |---------|---------------|-----:|-----------:| @@ -80,7 +79,6 @@ ON Year USING sum(Population); ``` -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -114,7 +112,6 @@ USING sum(Population) GROUP BY Country; ``` -
| Country | 2000 | 2010 | 2020 | |---------|-----:|-----:|-----:| @@ -133,7 +130,6 @@ USING sum(Population) GROUP BY Country; ``` -
| Country | 2000 | 2010 | |---------|-----:|-----:| @@ -158,7 +154,6 @@ ON Country, Name USING sum(Population); ``` -
| Year | NL_Amsterdam | NL_New York City | NL_Seattle | US_Amsterdam | US_New York City | US_Seattle | |-----:|-------------:|------------------|------------|--------------|-----------------:|-----------:| @@ -177,7 +172,6 @@ In this case, concatenating with an underscore is used to imitate the naming con PIVOT Cities ON Country || '_' || Name USING sum(Population); ``` -
| Year | NL_Amsterdam | US_New York City | US_Seattle | |-----:|-------------:|-----------------:|-----------:| @@ -200,7 +194,6 @@ USING sum(Population) AS total, max(Population) AS max GROUP BY Country; ``` -
| Country | 2000_total | 2000_max | 2010_total | 2010_max | 2020_total | 2020_max | |---------|-----------:|---------:|-----------:|---------:|-----------:|---------:| @@ -219,7 +212,6 @@ USING sum(Population) GROUP BY Country, Name; ``` -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -269,7 +261,6 @@ JOIN (PIVOT Cities ON Name USING sum(Population) GROUP BY Country) name_pivot USING (Country); ``` -
| Country | 2000 | 2010 | 2020 | Amsterdam | New York City | Seattle | |---------|-----:|-----:|-----:|----------:|--------------:|--------:| @@ -326,7 +317,6 @@ GROUP BY ALL; This produces the result: -
| Country | Name | list("YEAR") | list(population_sum) | |---------|---------------|--------------------|----------------------| @@ -336,7 +326,6 @@ This produces the result: The `PhysicalPivot` operator converts those lists into column names and values to return this result: -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -384,7 +373,6 @@ PIVOT ( ); ``` -
| Country | 2000 | 2010 | 2020 | |---------|-----:|-----:|-----:| diff --git a/docs/archive/0.10/sql/statements/unpivot.md b/docs/archive/0.10/sql/statements/unpivot.md index 4bac0f8c731..c884cf88d77 100644 --- a/docs/archive/0.10/sql/statements/unpivot.md +++ b/docs/archive/0.10/sql/statements/unpivot.md @@ -46,7 +46,6 @@ INSERT INTO monthly_sales VALUES FROM monthly_sales; ``` -
| empid | dept | Jan | Feb | Mar | Apr | May | Jun | |------:|-------------|----:|----:|----:|----:|----:|----:| @@ -72,7 +71,6 @@ INTO VALUE sales; ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -111,7 +109,6 @@ INTO VALUE sales; ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -153,7 +150,6 @@ UNPIVOT monthly_sales VALUE month_1_sales, month_2_sales, month_3_sales; ``` -
| empid | dept | quarter | month_1_sales | month_2_sales | month_3_sales | |------:|-------------|---------|--------------:|--------------:|--------------:| @@ -243,7 +239,6 @@ FROM monthly_sales; Note the single quotes to build a list of text strings to populate `month`, and the double quotes to pull the column values for use in `sales`. This produces the same result as the initial example: -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -297,7 +292,6 @@ FROM monthly_sales UNPIVOT ( ); ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -356,7 +350,6 @@ UNPIVOT ( ); ``` -
| empid | dept | quarter | month_1_sales | month_2_sales | month_3_sales | |------:|-------------|---------|--------------:|--------------:|--------------:| diff --git a/docs/archive/0.10/sql/statements/update.md b/docs/archive/0.10/sql/statements/update.md index 564630f828e..85bbd2eacce 100644 --- a/docs/archive/0.10/sql/statements/update.md +++ b/docs/archive/0.10/sql/statements/update.md @@ -48,7 +48,6 @@ SELECT * FROM original; ``` -
| key | value | |-----|------------------| @@ -80,7 +79,6 @@ SELECT * FROM original; ``` -
| key | value | |-----|-------------| diff --git a/docs/archive/0.10/sql/window_functions.md b/docs/archive/0.10/sql/window_functions.md index 4571d0dc959..c10dbca3748 100644 --- a/docs/archive/0.10/sql/window_functions.md +++ b/docs/archive/0.10/sql/window_functions.md @@ -234,7 +234,6 @@ CREATE TABLE "Generation History" AS After partitioning by plant and ordering by date, it will have this layout: -
| Plant | Date | MWh | |:---|:---|---:| @@ -280,7 +279,6 @@ ORDER BY 1, 2; The result will be the following: -
| Plant | Date | Row | |:---|:---|---:| @@ -352,7 +350,6 @@ and uses a `RANGE` frame of three days on either side of each day for the `avg` (to handle any missing days). This is the result: -
| Plant | Date | MWh 7-day Moving Average | |:---|:---|---:| diff --git a/docs/archive/0.9/api/adbc.md b/docs/archive/0.9/api/adbc.md index bfc09ed7062..cec84c4e1ae 100644 --- a/docs/archive/0.9/api/adbc.md +++ b/docs/archive/0.9/api/adbc.md @@ -82,7 +82,6 @@ Functions related to query execution: Functions related to binding, used for bulk insertion or in prepared statements. -
| Function Name | Description | Arguments | Example | |:---|:-|:---|:----| diff --git a/docs/archive/0.9/api/cli.md b/docs/archive/0.9/api/cli.md index 5ee5f76628e..807461c9f47 100644 --- a/docs/archive/0.9/api/cli.md +++ b/docs/archive/0.9/api/cli.md @@ -47,7 +47,6 @@ Once the CLI has been opened, enter a SQL statement followed by a semicolon, the SELECT 'quack' AS my_column; ``` -
| my_column | |-----------| @@ -59,7 +58,6 @@ SELECT 'excited quacking' AS another_column; ``` -
| my_column | another_column | |------------------------|------------------| @@ -186,7 +184,6 @@ SELECT student_id FROM 'data/ -> data/grades.csv The `.mode` command may be used to change the appearance of the tables returned in the terminal output. In addition to customizing the appearance, these modes have additional benefits. This can be useful for presenting DuckDB output elsewhere by redirecting the terminal output to a file, for example (see "Writing Results to a File" section below). Using the `insert` mode will build a series of SQL statements that can be used to insert the data at a later point. The `markdown` mode is particularly useful for building documentation! -
| mode | description | |------------|----------------------------------------------| diff --git a/docs/archive/0.9/api/python/types.md b/docs/archive/0.9/api/python/types.md index 7dac2e7b8f9..555c005dd70 100644 --- a/docs/archive/0.9/api/python/types.md +++ b/docs/archive/0.9/api/python/types.md @@ -18,7 +18,6 @@ This means that wherever a DuckDBPyType object is expected, it is also possible The table below shows the mapping of Python Builtin type to DuckDB type. -
| Type | DuckDB Type | |:----------------------------------------------|:-------------------| @@ -33,7 +32,6 @@ The table below shows the mapping of Python Builtin type to DuckDB type. The table below shows the mapping of Numpy DType to DuckDB type. -
| Type | DuckDB Type | |:----------------------------------------------|:-------------------| @@ -98,7 +96,6 @@ duckdb.typing.DuckDBPyType(Union[int, str, bool, bytearray]) For the builtin types, you can use the constants defined in `duckdb.typing` -
| DuckDB Type | |:-------------------| diff --git a/docs/archive/0.9/data/csv/auto_detection.md b/docs/archive/0.9/data/csv/auto_detection.md index 92de137d5df..37312756da6 100644 --- a/docs/archive/0.9/data/csv/auto_detection.md +++ b/docs/archive/0.9/data/csv/auto_detection.md @@ -26,7 +26,6 @@ Dialect detection works by attempting to parse the samples using the set of cons The following dialects are considered for automatic dialect detection. -
| Parameters | Considered values | |------------|-----------------------| @@ -58,7 +57,6 @@ After detecting the dialect, the system will attempt to figure out the types of The type detection works by attempting to convert the values in each column to the candidate types. If the conversion is unsuccessful, the candidate type is removed from the set of candidate types for that column. After all samples have been handled - the remaining candidate type with the highest priority is chosen. The set of considered candidate types in order of priority is given below: -
| Types | |-------------| @@ -94,7 +92,6 @@ If the ambiguities cannot be resolved by looking at the data the system has a li The system considers the following formats for dates (`dateformat`). Higher entries are chosen over lower entries in case of ambiguities (i.e., ISO 8601 is preferred over `MM-DD-YYYY`). -
| dateformat | |------------| @@ -108,7 +105,6 @@ The system considers the following formats for dates (`dateformat`). Higher entr The system considers the following formats for timestamps (`timestampformat`). Higher entries are chosen over lower entries in case of ambiguities. -
| timestampformat | |------------------------| diff --git a/docs/archive/0.9/data/csv/overview.md b/docs/archive/0.9/data/csv/overview.md index 522ca766b93..c96518b6736 100644 --- a/docs/archive/0.9/data/csv/overview.md +++ b/docs/archive/0.9/data/csv/overview.md @@ -83,7 +83,6 @@ The `read_csv_auto` is the simplest method of loading CSV files: it automaticall SELECT * FROM read_csv_auto('flights.csv'); ``` -
|FlightDate|UniqueCarrier| OriginCityName | DestCityName | |----------|-------------|-----------------|---------------| @@ -100,7 +99,6 @@ CREATE TABLE ontime AS SELECT * FROM read_csv_auto('flights.csv'); DESCRIBE ontime; ``` -
|Field |Type |Null|Key |Default|Extra| |--------------|-------|----|----|-------|-----| @@ -135,7 +133,6 @@ COPY ontime FROM 'flights.csv' (DELIMITER '|', HEADER); SELECT * FROM ontime; ``` -
|flightdate|uniquecarrier| origincityname | destcityname | |----------|-------------|-----------------|---------------| diff --git a/docs/archive/0.9/data/json/overview.md b/docs/archive/0.9/data/json/overview.md index 52cad3a9bd8..9a678afa6a9 100644 --- a/docs/archive/0.9/data/json/overview.md +++ b/docs/archive/0.9/data/json/overview.md @@ -92,7 +92,6 @@ Each line is a JSON. SELECT * FROM read_json_auto('records.json', format='newline_delimited'); ``` -
| key1 | key2 | |----------|----------| @@ -114,7 +113,6 @@ If the JSON file contains a JSON array of objects (pretty-printed or not), `arra SELECT * FROM read_json_auto('array.json', format='array'); ``` -
| key1 | key2 | |----------|----------| @@ -143,7 +141,6 @@ If the JSON file contains JSON that is not newline-delimited or an array, `unstr SELECT * FROM read_json_auto('unstructured.json', format='unstructured'); ``` -
| key1 | key2 | |----------|----------| @@ -166,7 +163,6 @@ Continuing with the same example file from before: SELECT * FROM read_json_auto('records.json', records=true); ``` -
| key1 | key2 | |----------|----------| @@ -179,7 +175,6 @@ When `records=false`, the JSON extension will not unpack the top-level objects, SELECT * FROM read_json_auto('records.json', records=false); ``` -
| json | |------------------------------------| @@ -197,7 +192,6 @@ This is especially useful if we have non-object JSON, for example: SELECT * FROM read_json_auto('arrays.json', records=false); ``` -
| json | |-------------| @@ -217,7 +211,6 @@ The `read_json_auto` is the simplest method of loading JSON files: it automatica SELECT * FROM read_json_auto('todos.json') LIMIT 5; ``` -
| userId | id | title | completed | |--------|----|-----------------------------------------------------------------|-----------| @@ -236,7 +229,6 @@ CREATE TABLE todos AS SELECT * FROM read_json_auto('todos.json'); DESCRIBE todos; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|-----|---------|-------| @@ -267,7 +259,6 @@ COPY todos FROM 'todos.json'; SELECT * FROM todos LIMIT 5; ``` -
| userId | id | title | completed | |--------|----|-----------------------------------------------------------------|-----------| diff --git a/docs/archive/0.9/data/multiple_files/combining_schemas.md b/docs/archive/0.9/data/multiple_files/combining_schemas.md index 34f0ff93b44..8e769e8236f 100644 --- a/docs/archive/0.9/data/multiple_files/combining_schemas.md +++ b/docs/archive/0.9/data/multiple_files/combining_schemas.md @@ -47,7 +47,6 @@ FlightDate|UniqueCarrier|OriginCityName|DestCityName Reading the two files at the same time will produce the following result set: -
| FlightDate | UniqueCarrier | OriginCityName | DestCityName | |------------|---------------|----------------|-----------------| @@ -82,7 +81,6 @@ Reading these when unifying column names **by position** results in an error - a SELECT * FROM read_csv_auto(['flights3.csv', 'flights4.csv'], union_by_name=true); ``` -
| FlightDate | OriginCityName | DestCityName | UniqueCarrier | |------------|----------------|-----------------|---------------| diff --git a/docs/archive/0.9/data/multiple_files/overview.md b/docs/archive/0.9/data/multiple_files/overview.md index 7bf7023ef1f..ff1f86d2d7b 100644 --- a/docs/archive/0.9/data/multiple_files/overview.md +++ b/docs/archive/0.9/data/multiple_files/overview.md @@ -59,7 +59,6 @@ SELECT * FROM read_parquet(['file1.parquet', 'file2.parquet', 'file3.parquet']); Any file name input to the read_parquet function can either be an exact filename, or use a glob syntax to read multiple files that match a pattern. -
| Wildcard | Description | |------------|-----------------------------------------------------------| @@ -97,7 +96,6 @@ The `filename` argument can be used to add an extra `filename` column to the res SELECT * FROM read_csv_auto(['flights1.csv', 'flights2.csv'], union_by_name=true, filename=true); ``` -
| FlightDate | OriginCityName | DestCityName | UniqueCarrier | filename | |------------|----------------|-----------------|---------------|--------------| @@ -116,7 +114,6 @@ It accepts one parameter: the path to search (which may include glob patterns). SELECT * FROM glob('*'); ``` -
| file | |---------------| diff --git a/docs/archive/0.9/data/parquet/metadata.md b/docs/archive/0.9/data/parquet/metadata.md index 25f36709cac..14efdb94982 100644 --- a/docs/archive/0.9/data/parquet/metadata.md +++ b/docs/archive/0.9/data/parquet/metadata.md @@ -17,7 +17,6 @@ SELECT * FROM parquet_metadata('test.parquet'); Below is a table of the columns returned by `parquet_metadata`. -
| Field | Type | |---------------------------|-----------| @@ -59,7 +58,6 @@ SELECT * FROM parquet_schema('test.parquet'); Below is a table of the columns returned by `parquet_schema`. -
| Field | Type | |-------------------|-----------| diff --git a/docs/archive/0.9/extensions/autocomplete.md b/docs/archive/0.9/extensions/autocomplete.md index df30793613e..50f71791d98 100644 --- a/docs/archive/0.9/extensions/autocomplete.md +++ b/docs/archive/0.9/extensions/autocomplete.md @@ -9,7 +9,6 @@ title: AutoComplete Extension This extension adds supports for autocomplete in the [CLI client](../api/cli). -
| Function | Description | |:----------------------------------------|:-------------------------------------------------------| @@ -23,7 +22,6 @@ SELECT * FROM sql_auto_complete('SEL'); Returns: -
| suggestion | suggestion_start | |-------------|------------------| diff --git a/docs/archive/0.9/extensions/full_text_search.md b/docs/archive/0.9/extensions/full_text_search.md index 682f4f8a863..45d651eaa24 100644 --- a/docs/archive/0.9/extensions/full_text_search.md +++ b/docs/archive/0.9/extensions/full_text_search.md @@ -43,7 +43,6 @@ drop_fts_index(input_table) Drops a FTS index for the specified table. -
| Name | Type | Description | |:--|:--|:-----------| diff --git a/docs/archive/0.9/extensions/httpfs.md b/docs/archive/0.9/extensions/httpfs.md index 4cd145bbd2b..a41f3b456de 100644 --- a/docs/archive/0.9/extensions/httpfs.md +++ b/docs/archive/0.9/extensions/httpfs.md @@ -49,7 +49,6 @@ The `httpfs` extension supports reading/writing/globbing files on object storage The `httpfs` filesystem is tested with [AWS S3](https://aws.amazon.com/s3/), [Minio](https://min.io/), [Google Cloud](https://cloud.google.com/storage/docs/interoperability), and [lakeFS](https://docs.lakefs.io/integrations/duckdb.html). Other services that implement the S3 API should also work, but not all features may be supported. Below is a list of which parts of the S3 API are required for each `httpfs` feature. -
| Feature | Required S3 API features | |:---|:---| @@ -156,7 +155,6 @@ SELECT * FROM read_parquet('s3://bucket/*.parquet', FILENAME = 1); could for example result in: -
| column_a | column_b | filename | |:---|:---|:---| @@ -181,7 +179,6 @@ SELECT * FROM read_parquet('s3://bucket/*/file.parquet', HIVE_PARTITIONING = 1); could result in: -
| column_a | column_b | year | |:---|:---|:---| @@ -225,7 +222,6 @@ s3://my-bucket/partitioned/part_col_a=/part_col_b=/data_ | setting | description | |:---|:---| diff --git a/docs/archive/0.9/extensions/mysql.md b/docs/archive/0.9/extensions/mysql.md index a7b24f23f0c..74d9b74798b 100644 --- a/docs/archive/0.9/extensions/mysql.md +++ b/docs/archive/0.9/extensions/mysql.md @@ -37,7 +37,6 @@ USE mysqldb; The connection string determines the parameters for how to connect to MySQL as a set of `key=value` pairs. Any options not provided are replaced by their default values, as per the table below. -
| Setting | Default | |----------|--------------| diff --git a/docs/archive/0.9/extensions/spatial.md b/docs/archive/0.9/extensions/spatial.md index 4516e3ab2b2..4ed5233b1e0 100644 --- a/docs/archive/0.9/extensions/spatial.md +++ b/docs/archive/0.9/extensions/spatial.md @@ -142,7 +142,6 @@ Compute relationships and spatial predicates between geometries. ## Spatial Aggregate Functions -
| Aggregate functions | Implemented with | |-------------------------------------------|------------------| diff --git a/docs/archive/0.9/guides/meta/list_tables.md b/docs/archive/0.9/guides/meta/list_tables.md index a3af89231e2..79799e798c7 100644 --- a/docs/archive/0.9/guides/meta/list_tables.md +++ b/docs/archive/0.9/guides/meta/list_tables.md @@ -14,7 +14,6 @@ CREATE TABLE tbl(i INTEGER); SHOW TABLES; ``` -
| name | |------| @@ -29,7 +28,6 @@ CREATE TABLE s1.tbl(v VARCHAR); SHOW ALL TABLES; ``` -
| database | schema | table_name | column_names | column_types | temporary | |----------|--------|------------|--------------|--------------|-----------| @@ -43,7 +41,6 @@ CREATE TABLE tbl(i INTEGER PRIMARY KEY, j VARCHAR); DESCRIBE tbl; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|------|---------|-------| diff --git a/docs/archive/0.9/sql/data_types/bitstring.md b/docs/archive/0.9/sql/data_types/bitstring.md index a18ebe9986c..66b594da38a 100644 --- a/docs/archive/0.9/sql/data_types/bitstring.md +++ b/docs/archive/0.9/sql/data_types/bitstring.md @@ -8,7 +8,6 @@ redirect_from: title: Bitstring Type --- -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/0.9/sql/data_types/blob.md b/docs/archive/0.9/sql/data_types/blob.md index 0302bd7c1bd..5fa5e4e49b9 100644 --- a/docs/archive/0.9/sql/data_types/blob.md +++ b/docs/archive/0.9/sql/data_types/blob.md @@ -9,7 +9,6 @@ redirect_from: title: Blob Type --- -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/0.9/sql/data_types/boolean.md b/docs/archive/0.9/sql/data_types/boolean.md index 645a67546ab..d900da65489 100644 --- a/docs/archive/0.9/sql/data_types/boolean.md +++ b/docs/archive/0.9/sql/data_types/boolean.md @@ -8,7 +8,6 @@ redirect_from: title: Boolean Type --- -
| Name | Aliases | Description | |:---|:---|:---| @@ -42,7 +41,6 @@ The `AND`/`OR` conjunctions can be used to combine boolean values. Below is the truth table for the `AND` conjunction (i.e., `x AND y`). -
| X | X AND true | X AND false | X AND NULL | |-------|-------|-------|-------| @@ -52,7 +50,6 @@ Below is the truth table for the `AND` conjunction (i.e., `x AND y`). Below is the truth table for the `OR` conjunction (i.e., `x OR y`). -
| X | X OR true | X OR false | X OR NULL | |-------|------|-------|------| diff --git a/docs/archive/0.9/sql/data_types/date.md b/docs/archive/0.9/sql/data_types/date.md index 9441b0f62a7..73e83b9c855 100644 --- a/docs/archive/0.9/sql/data_types/date.md +++ b/docs/archive/0.9/sql/data_types/date.md @@ -8,7 +8,6 @@ redirect_from: title: Date Types --- -
| Name | Aliases | Description | |:-------|:--------|:--------------------------------| @@ -25,7 +24,6 @@ SELECT DATE '1992-09-20'; There are also three special date values that can be used on input: -
| Input String | Description | |:-------------|:----------------------------------| @@ -40,7 +38,6 @@ but `epoch` is simply a notational shorthand that will be converted to the date SELECT '-infinity'::DATE, 'epoch'::DATE, 'infinity'::DATE; ``` -
| Negative | Epoch | Positive | |:----------|:-----------|:---------| diff --git a/docs/archive/0.9/sql/data_types/enum.md b/docs/archive/0.9/sql/data_types/enum.md index 34ec221ad14..291fed36330 100644 --- a/docs/archive/0.9/sql/data_types/enum.md +++ b/docs/archive/0.9/sql/data_types/enum.md @@ -9,7 +9,6 @@ redirect_from: title: Enum Types --- -
| Name | Description | |:--|:-----| @@ -58,7 +57,6 @@ CREATE TYPE birds AS ENUM (SELECT my_varchar FROM my_inputs); SELECT enum_range(NULL::birds) AS my_enum_range; ``` -
| my_enum_range | |-----------------| diff --git a/docs/archive/0.9/sql/data_types/interval.md b/docs/archive/0.9/sql/data_types/interval.md index 159747a1970..998a5210af6 100644 --- a/docs/archive/0.9/sql/data_types/interval.md +++ b/docs/archive/0.9/sql/data_types/interval.md @@ -11,7 +11,6 @@ title: Interval Type Intervals represent a period of time. This period can be measured in a specific unit or combination of units, for example years, days, or seconds. Intervals are generally used to *modify* timestamps or dates by either adding or subtracting them. -
| Name | Description | |:---|:---| diff --git a/docs/archive/0.9/sql/data_types/numeric.md b/docs/archive/0.9/sql/data_types/numeric.md index 0e1eb08324c..b7e2032f258 100644 --- a/docs/archive/0.9/sql/data_types/numeric.md +++ b/docs/archive/0.9/sql/data_types/numeric.md @@ -33,7 +33,6 @@ The data type `DECIMAL(WIDTH, SCALE)` represents an exact fixed-point decimal va Internally, decimals are represented as integers depending on their specified width. -
| Width | Internal | Size (Bytes) | |:---|:---|---:| @@ -48,7 +47,6 @@ Performance can be impacted by using too large decimals when not required. In pa The data types `REAL` and `DOUBLE` precision are inexact, variable-precision numeric types. In practice, these types are usually implementations of IEEE Standard 754 for Binary Floating-Point Arithmetic (single and double precision, respectively), to the extent that the underlying processor, operating system, and compiler support it. -
| Name | Aliases | Description | |:--|:--|:--------| diff --git a/docs/archive/0.9/sql/data_types/struct.md b/docs/archive/0.9/sql/data_types/struct.md index 0819674453b..e354e81be03 100644 --- a/docs/archive/0.9/sql/data_types/struct.md +++ b/docs/archive/0.9/sql/data_types/struct.md @@ -83,7 +83,6 @@ This is particularly useful when a prior operation creates a struct of unknown s SELECT a.* FROM (SELECT {'x':1, 'y':2, 'z':3} AS a); ``` -
| x | y | z | |:---|:---|:---| @@ -140,7 +139,6 @@ This makes it not a good candidate to use directly in the result of a query, `st #### Example Data Table Named t1 -
| my_column | another_column | |:---|:---| @@ -158,7 +156,6 @@ FROM t1; #### Example Output -
| my_struct_column | identical_struct_column | |:---|:---| @@ -178,7 +175,6 @@ FROM t1; #### Example Output -
| my_struct_column | identical_struct_column | |:---|:---| diff --git a/docs/archive/0.9/sql/data_types/text.md b/docs/archive/0.9/sql/data_types/text.md index 3176d025f4e..a7c25f713dc 100644 --- a/docs/archive/0.9/sql/data_types/text.md +++ b/docs/archive/0.9/sql/data_types/text.md @@ -10,7 +10,6 @@ title: Text Types In DuckDB, strings can be stored in the `VARCHAR` field. -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/0.9/sql/data_types/timestamp.md b/docs/archive/0.9/sql/data_types/timestamp.md index 2487db276f3..8e8401561f5 100644 --- a/docs/archive/0.9/sql/data_types/timestamp.md +++ b/docs/archive/0.9/sql/data_types/timestamp.md @@ -38,7 +38,6 @@ SELECT TIMESTAMP WITH TIME ZONE '1992-09-20 11:30:00.123456'; There are also three special date values that can be used on input: -
| Input String | Valid Types | Description | |:-------------|:--------------------------------------|:-----------------------------------------------| @@ -53,7 +52,6 @@ but `epoch` is simply a notational shorthand that will be converted to the time SELECT '-infinity'::TIMESTAMP, 'epoch'::TIMESTAMP, 'infinity'::TIMESTAMP; ``` -
| Negative | Epoch | Positive | |:----------|:--------------------|:---------| diff --git a/docs/archive/0.9/sql/data_types/timezones.md b/docs/archive/0.9/sql/data_types/timezones.md index d9e44664b0e..2506e69812d 100644 --- a/docs/archive/0.9/sql/data_types/timezones.md +++ b/docs/archive/0.9/sql/data_types/timezones.md @@ -22,7 +22,6 @@ ORDER BY name; ``` -
| name | abbrev | utc_offset | |----------------------------------|----------------------------------|-----------:| diff --git a/docs/archive/0.9/sql/data_types/union.md b/docs/archive/0.9/sql/data_types/union.md index 0df1770ea4c..64f0e6609d2 100644 --- a/docs/archive/0.9/sql/data_types/union.md +++ b/docs/archive/0.9/sql/data_types/union.md @@ -74,7 +74,6 @@ So how do we disambiguate if we want to create a `UNION` with multiple members o `UNION` types can be cast between each other if the source type is a "subset" of the target type. In other words, all the tags in the source `UNION` must be present in the target `UNION`, and all the types of the matching tags must be implicitly castable between source and target. In essence, this means that `UNION` types are covariant with respect to their members. -
| Ok | Source | Target | Comments | |----|----------------------|-----------------------|------------------------------------| diff --git a/docs/archive/0.9/sql/duckdb_table_functions.md b/docs/archive/0.9/sql/duckdb_table_functions.md index c6f8a503a5c..e2265ce48a3 100644 --- a/docs/archive/0.9/sql/duckdb_table_functions.md +++ b/docs/archive/0.9/sql/duckdb_table_functions.md @@ -165,7 +165,6 @@ Note that `duckdb_indexes` only provides metadata about secondary indexes - i.e. The `duckdb_keywords()` function provides metadata about DuckDB's keywords and reserved words. -
| Column | Description | Type | |:-|:---|:-| @@ -218,7 +217,6 @@ Attributes like `temporary`, `start_value` etc. correspond to the various option The `duckdb_settings()` function provides metadata about the settings available in the DuckDB instance. -
| Column | Description | Type | |:-|:---|:-| @@ -292,7 +290,6 @@ The [`information_schema.tables`](information_schema#tables-and-views) system vi The `duckdb_temporary_files()` function provides metadata about the temporary files DuckDB has written to disk, to offload data from memory. This function mostly exists for debugging and testing purposes. -
| Column | Description | Type | |:-|:---|:-| diff --git a/docs/archive/0.9/sql/expressions/comparison_operators.md b/docs/archive/0.9/sql/expressions/comparison_operators.md index b62d22f0611..15074070b9b 100644 --- a/docs/archive/0.9/sql/expressions/comparison_operators.md +++ b/docs/archive/0.9/sql/expressions/comparison_operators.md @@ -15,7 +15,6 @@ title: Comparisons The table below shows the standard comparison operators. Whenever either of the input arguments is `NULL`, the output of the comparison is `NULL`. -
| Operator | Description | Example | Result | |:---|:---|:---|:---| @@ -29,7 +28,6 @@ Whenever either of the input arguments is `NULL`, the output of the comparison i The table below shows the standard distinction operators. These operators treat `NULL` values as equal. -
| Operator | Description | Example | Result | |:---|:---|:---|:-| @@ -43,7 +41,6 @@ These operators treat `NULL` values as equal. Besides the standard comparison operators there are also the `BETWEEN` and `IS (NOT) NULL` operators. These behave much like operators, but have special syntax mandated by the SQL standard. They are shown in the table below. Note that BETWEEN and NOT BETWEEN are only equivalent to the examples below in the cases where both `a`, `x` and `y` are of the same type, as BETWEEN will cast all of its inputs to the same type. -
| Predicate | Description | |:---|:---| diff --git a/docs/archive/0.9/sql/expressions/logical_operators.md b/docs/archive/0.9/sql/expressions/logical_operators.md index daa20a75a54..331616c9bb6 100644 --- a/docs/archive/0.9/sql/expressions/logical_operators.md +++ b/docs/archive/0.9/sql/expressions/logical_operators.md @@ -12,7 +12,6 @@ title: Logical Operators The following logical operators are available: `AND`, `OR` and `NOT`. SQL uses a three-valuad logic system with `true`, `false` and `NULL`. Note that logical operators involving `NULL` do not always evaluate to `NULL`. For example, `NULL AND false` will evaluate to `false`, and `NULL OR true` will evaluate to `true`. Below are the complete truth tables: -
| `a` | `b` | `a AND b` | `a OR b` | |:---|:---|:---|:---| @@ -23,7 +22,6 @@ The following logical operators are available: `AND`, `OR` and `NOT`. SQL uses a | `false` | `NULL` | `false` | `NULL` | | `NULL` | `NULL` | `NULL` | `NULL`| -
| `a` | `NOT a` | |:---|:---| diff --git a/docs/archive/0.9/sql/expressions/star.md b/docs/archive/0.9/sql/expressions/star.md index 017d42644c8..284bfdd47b4 100644 --- a/docs/archive/0.9/sql/expressions/star.md +++ b/docs/archive/0.9/sql/expressions/star.md @@ -65,7 +65,6 @@ INSERT INTO numbers VALUES (1, 10), (2, 20), (3, NULL); SELECT MIN(COLUMNS(*)), COUNT(COLUMNS(*)) FROM numbers; ``` -
| min(numbers.id) | min(numbers.number) | count(numbers.id) | count(numbers.number) | |-----------------|---------------------|-------------------|-----------------------| @@ -77,7 +76,6 @@ The `*` expression in the `COLUMNS` statement can also contain `EXCLUDE` or `REP SELECT MIN(COLUMNS(* REPLACE (number + id AS number))), COUNT(COLUMNS(* EXCLUDE (number))) FROM numbers; ``` -
| min(numbers.id) | min(number := (number + id)) | count(numbers.id) | |-----------------|------------------------------|-------------------| @@ -89,7 +87,6 @@ COLUMNS expressions can also be combined, as long as the `COLUMNS` contains the SELECT COLUMNS(*) + COLUMNS(*) FROM numbers; ``` -
| (numbers.id + numbers.id) | (numbers.number + numbers.number) | |---------------------------|-----------------------------------| @@ -106,7 +103,6 @@ SELECT COLUMNS(*) + COLUMNS(*) FROM numbers; SELECT COLUMNS('(id|numbers?)') FROM numbers; ``` -
| id | number | |----|--------| @@ -122,7 +118,6 @@ SELECT COLUMNS('(id|numbers?)') FROM numbers; SELECT COLUMNS(c -> c LIKE '%num%') FROM numbers; ``` -
| number | |--------| @@ -141,7 +136,6 @@ See the [struct](../data_types/struct) and [nested function](../functions/nested SELECT a.* FROM (SELECT {'x':1, 'y':2, 'z':3} AS a); ``` -
| x | y | z | |:---|:---|:---| diff --git a/docs/archive/0.9/sql/expressions/subqueries.md b/docs/archive/0.9/sql/expressions/subqueries.md index 1f20e238d00..b14bb83792e 100644 --- a/docs/archive/0.9/sql/expressions/subqueries.md +++ b/docs/archive/0.9/sql/expressions/subqueries.md @@ -18,7 +18,6 @@ Consider the following table: ### Grades -
| grade | course | |---:|:---| diff --git a/docs/archive/0.9/sql/functions/bitstring.md b/docs/archive/0.9/sql/functions/bitstring.md index 31961455720..69781d052ec 100644 --- a/docs/archive/0.9/sql/functions/bitstring.md +++ b/docs/archive/0.9/sql/functions/bitstring.md @@ -14,7 +14,6 @@ Bitstrings must be of equal length when performing the bitwise operands AND, OR The table below shows the available mathematical operators for `BIT` type. -
| Operator | Description | Example | Result | |:---|:---|:---|:---| diff --git a/docs/archive/0.9/sql/functions/dateformat.md b/docs/archive/0.9/sql/functions/dateformat.md index 7b5cf11da39..70350f89923 100644 --- a/docs/archive/0.9/sql/functions/dateformat.md +++ b/docs/archive/0.9/sql/functions/dateformat.md @@ -47,7 +47,6 @@ SELECT * FROM read_csv('test.csv', dateformat='%m/%d/%Y'); Below is a full list of all available format specifiers. -
| Specifier | Description | Example | |:-|:------|:---| diff --git a/docs/archive/0.9/sql/functions/numeric.md b/docs/archive/0.9/sql/functions/numeric.md index 4c2be4ebad2..0207a856f8a 100644 --- a/docs/archive/0.9/sql/functions/numeric.md +++ b/docs/archive/0.9/sql/functions/numeric.md @@ -11,7 +11,6 @@ title: Numeric Functions The table below shows the available mathematical operators for numeric types. -
| Operator | Description | Example | Result | |:-|:---|:-|:-| diff --git a/docs/archive/0.9/sql/functions/patternmatching.md b/docs/archive/0.9/sql/functions/patternmatching.md index 709c7e23a11..a8b71afee29 100644 --- a/docs/archive/0.9/sql/functions/patternmatching.md +++ b/docs/archive/0.9/sql/functions/patternmatching.md @@ -54,7 +54,6 @@ Additionally, the function `like_escape` has the same functionality as a `LIKE` There are also alternative characters that can be used as keywords in place of `LIKE` expressions. These enhance PostgreSQL compatibility. -
| LIKE-style | PostgreSQL-style | |:---|:---| @@ -84,7 +83,6 @@ Some examples: There are also alternative characters that can be used as keywords in place of `SIMILAR TO` expressions. These follow POSIX syntax. -
| SIMILAR TO-style | POSIX-style | |:---|:---| @@ -120,7 +118,6 @@ NOT 'best.txt' GLOB '*.txt' -- false Three tildes (`~~~`) may also be used in place of the `GLOB` keyword. -
| GLOB-style | Symbolic-style | |:---|:---| @@ -136,7 +133,6 @@ It accepts one parameter: the path to search (which may include glob patterns). SELECT * FROM glob('*'); ``` -
| file | |---------------| @@ -177,7 +173,6 @@ regexp_matches('abc', '(?i)A') -- true The `regexp_matches` function also supports the following options. -
| Option | Description | |:---|:---| @@ -198,7 +193,6 @@ regexp_matches('hello\nworld', 'hello.world', 's') -- true The `regexp_matches` operator will be optimized to the `LIKE` operator when possible. To achieve the best results, the `'s'` option should be passed. By default the `RE2` library doesn't match '.' to newline. -
| Original | Optimized equivalent | |:---|:---| diff --git a/docs/archive/0.9/sql/functions/time.md b/docs/archive/0.9/sql/functions/time.md index 2797ef90013..f2627f82c4d 100644 --- a/docs/archive/0.9/sql/functions/time.md +++ b/docs/archive/0.9/sql/functions/time.md @@ -13,7 +13,6 @@ This section describes functions and operators for examining and manipulating `T The table below shows the available mathematical operators for `TIME` types. -
| Operator | Description | Example | Result | |:-|:---|:----|:--| diff --git a/docs/archive/0.9/sql/functions/utility.md b/docs/archive/0.9/sql/functions/utility.md index 6a0a08d6891..dd0801fa639 100644 --- a/docs/archive/0.9/sql/functions/utility.md +++ b/docs/archive/0.9/sql/functions/utility.md @@ -41,7 +41,6 @@ The functions below are difficult to categorize into specific function types and A table function is used in place of a table in a `FROM` clause. -
| Function | Description | Example | |:--|:---|:-| diff --git a/docs/archive/0.9/sql/indexes.md b/docs/archive/0.9/sql/indexes.md index a8b5686f45d..692fdb04a66 100644 --- a/docs/archive/0.9/sql/indexes.md +++ b/docs/archive/0.9/sql/indexes.md @@ -38,7 +38,6 @@ Both min-max indexes and ART indexes are persisted on disk. ### Parameters -
| Name | Description | |:-|:-----| @@ -71,7 +70,6 @@ CREATE INDEX i_index ON integers ((j+k)); ### Parameters -
| Name | Description | |:---|:---| diff --git a/docs/archive/0.9/sql/pragmas.md b/docs/archive/0.9/sql/pragmas.md index eb385f01707..865ed60eee7 100644 --- a/docs/archive/0.9/sql/pragmas.md +++ b/docs/archive/0.9/sql/pragmas.md @@ -268,7 +268,6 @@ CALL pragma_storage_info('table_name'); This call returns the following information for the given table: -
| name | type | description | |----------------|-----------|-------------------------------------------------------| diff --git a/docs/archive/0.9/sql/query_syntax/filter.md b/docs/archive/0.9/sql/query_syntax/filter.md index 5e0e09b95e2..5bf1fda0f23 100644 --- a/docs/archive/0.9/sql/query_syntax/filter.md +++ b/docs/archive/0.9/sql/query_syntax/filter.md @@ -27,7 +27,6 @@ SELECT FROM generate_series(1, 10) tbl(i); ``` -
| total_rows | lte_five | odds | |:---|:---|:---| @@ -44,7 +43,6 @@ SELECT FROM generate_series(1, 10) tbl(i); ``` -
| lte_five_sum | odds_median | odds_lte_five_median | |:---|:---|:---| @@ -89,7 +87,6 @@ SELECT FROM stacked_data; ``` -
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| @@ -108,7 +105,6 @@ SELECT FROM stacked_data; ``` -
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| @@ -125,7 +121,6 @@ SELECT FROM stacked_data; ``` -
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| diff --git a/docs/archive/0.9/sql/query_syntax/orderby.md b/docs/archive/0.9/sql/query_syntax/orderby.md index 6f4f54b1a13..2e7f5871cab 100644 --- a/docs/archive/0.9/sql/query_syntax/orderby.md +++ b/docs/archive/0.9/sql/query_syntax/orderby.md @@ -87,7 +87,6 @@ FROM addresses ORDER BY ALL; ``` -
| address | city | zip | |------------------------|-----------|------------| @@ -104,7 +103,6 @@ FROM addresses ORDER BY ALL DESC; ``` -
| address | city | zip | |------------------------|-----------|------------| diff --git a/docs/archive/0.9/sql/query_syntax/qualify.md b/docs/archive/0.9/sql/query_syntax/qualify.md index 4e438509adf..b93a0f46167 100644 --- a/docs/archive/0.9/sql/query_syntax/qualify.md +++ b/docs/archive/0.9/sql/query_syntax/qualify.md @@ -79,7 +79,6 @@ WHERE function_rank < 3; ``` -
| schema_name | function_name | function_rank | |:---|:---|:---| diff --git a/docs/archive/0.9/sql/statements/insert.md b/docs/archive/0.9/sql/statements/insert.md index dda457f7450..006022874b2 100644 --- a/docs/archive/0.9/sql/statements/insert.md +++ b/docs/archive/0.9/sql/statements/insert.md @@ -111,7 +111,6 @@ INSERT INTO t1 RETURNING *; ``` -
| i | |---| @@ -125,7 +124,6 @@ INSERT INTO t2 RETURNING *, i * j AS i_times_j; ``` -
| i | j | i_times_j | |---|---|---| @@ -143,7 +141,6 @@ INSERT INTO t3 RETURNING *; ``` -
| i | j | |---|---| diff --git a/docs/archive/0.9/sql/statements/pivot.md b/docs/archive/0.9/sql/statements/pivot.md index 7bfb20bc166..ecbe1eadfed 100644 --- a/docs/archive/0.9/sql/statements/pivot.md +++ b/docs/archive/0.9/sql/statements/pivot.md @@ -50,7 +50,6 @@ INSERT INTO Cities VALUES ('US', 'New York City', 2020, 8772); FROM Cities; ``` -
| Country | Name | Year | Population | |---------|---------------|------|------------| @@ -78,7 +77,6 @@ If the `USING` clause is not included, it defaults to `COUNT(*)`. PIVOT Cities ON Year USING SUM(Population); ``` -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|------|------|------| @@ -107,7 +105,6 @@ In the below example, the Name column is no longer included in the output, and t PIVOT Cities ON Year USING SUM(Population) GROUP BY Country; ``` -
| Country | 2000 | 2010 | 2020 | |---------|------|------|------| @@ -123,7 +120,6 @@ Let's say for example that we wanted to forget about the year 2020 for no partic PIVOT Cities ON Year IN (2000, 2010) USING SUM(Population) GROUP BY Country; ``` -
| Country | 2000 | 2010 | |---------|------|------| @@ -146,7 +142,6 @@ Some combinations may not be present in the underlying data, so those columns ar PIVOT Cities ON Country, Name USING SUM(Population); ``` -
| Year | NL_Amsterdam | NL_New York City | NL_Seattle | US_Amsterdam | US_New York City | US_Seattle | |------|--------------|------------------|------------|--------------|------------------|------------| @@ -164,7 +159,6 @@ In this case, concatenating with an underscore is used to imitate the naming con PIVOT Cities ON Country || '_' || Name USING SUM(Population); ``` -
| Year | NL_Amsterdam | US_New York City | US_Seattle | |------|--------------|------------------|------------| @@ -184,7 +178,6 @@ In this example, both the `SUM` and `MAX` of the Population column are calculate PIVOT Cities ON Year USING SUM(Population) AS total, MAX(Population) AS max GROUP BY Country; ``` -
| Country | 2000_total | 2000_max | 2010_total | 2010_max | 2020_total | 2020_max | |---------|------------|----------|------------|----------|------------|----------| @@ -201,7 +194,6 @@ Note that column names must be used rather than column positions (1, 2, etc.), a PIVOT Cities ON Year USING SUM(Population) GROUP BY Country, Name; ``` -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|------|------|------| @@ -246,7 +238,6 @@ JOIN USING (Country); ``` -
| Country | 2000 | 2010 | 2020 | Amsterdam | New York City | Seattle | |---------|------|------|------|-----------|---------------|---------| @@ -297,7 +288,6 @@ GROUP BY ALL; This produces the result: -
| Country | Name | list("YEAR") | list(population_sum) | |---------|---------------|--------------------|----------------------| @@ -307,7 +297,6 @@ This produces the result: The `PhysicalPivot` operator converts those lists into column names and values to return this result: -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|------|------|------| @@ -355,7 +344,6 @@ PIVOT ( ); ``` -
| Country | 2000 | 2010 | 2020 | |---------|------|------|------| diff --git a/docs/archive/0.9/sql/statements/unpivot.md b/docs/archive/0.9/sql/statements/unpivot.md index 68dea6b7abc..0b2cc56f61a 100644 --- a/docs/archive/0.9/sql/statements/unpivot.md +++ b/docs/archive/0.9/sql/statements/unpivot.md @@ -47,7 +47,6 @@ INSERT INTO monthly_sales VALUES FROM monthly_sales; ``` -
| empid | dept | Jan | Feb | Mar | Apr | May | Jun | |-------|-------------|-----|-----|-----|-----|-----|-----| @@ -73,7 +72,6 @@ INTO VALUE sales; ``` -
| empid | dept | month | sales | |-------|-------------|-------|-------| @@ -113,7 +111,6 @@ INTO VALUE sales; ``` -
| empid | dept | month | sales | |-------|-------------|-------|-------| @@ -156,7 +153,6 @@ UNPIVOT monthly_sales VALUE month_1_sales, month_2_sales, month_3_sales; ``` -
| empid | dept | quarter | month_1_sales | month_2_sales | month_3_sales | |-------|-------------|---------|---------------|---------------|---------------| @@ -228,7 +224,6 @@ FROM monthly_sales; Note the single quotes to build a list of text strings to populate `month`, and the double quotes to pull the column values for use in `sales`. This produces the same result as the initial example: -
| empid | dept | month | sales | |-------|-------------|-------|-------| @@ -285,7 +280,6 @@ FROM monthly_sales UNPIVOT ( ); ``` -
| empid | dept | month | sales | |-------|-------------|-------|-------| @@ -345,7 +339,6 @@ UNPIVOT ( ); ``` -
| empid | dept | quarter | month_1_sales | month_2_sales | month_3_sales | |-------|-------------|---------|---------------|---------------|---------------| diff --git a/docs/archive/0.9/sql/statements/update.md b/docs/archive/0.9/sql/statements/update.md index a9f427318c4..b06bccaa465 100644 --- a/docs/archive/0.9/sql/statements/update.md +++ b/docs/archive/0.9/sql/statements/update.md @@ -42,7 +42,6 @@ CREATE OR REPLACE TABLE new AS SELECT * FROM original; ``` -
| key | value | |-----|------------------| @@ -69,7 +68,6 @@ UPDATE original SELECT * FROM original; ``` -
| key | value | |-----|-------------| diff --git a/docs/archive/1.0/api/adbc.md b/docs/archive/1.0/api/adbc.md index a088943ee9f..d0b65d759a0 100644 --- a/docs/archive/1.0/api/adbc.md +++ b/docs/archive/1.0/api/adbc.md @@ -74,7 +74,6 @@ Functions related to query execution: Functions related to binding, used for bulk insertion or in prepared statements. -
| Function name | Description | Arguments | Example | |:---|:-|:---|:----| diff --git a/docs/archive/1.0/api/cli/arguments.md b/docs/archive/1.0/api/cli/arguments.md index dffe164901e..add3c91381e 100644 --- a/docs/archive/1.0/api/cli/arguments.md +++ b/docs/archive/1.0/api/cli/arguments.md @@ -12,7 +12,6 @@ duckdb -help For a list of dot commands available in the CLI shell, see the [Dot Commands page]({% link docs/archive/1.0/api/cli/dot_commands.md %}). -
diff --git a/docs/archive/1.0/api/cli/dot_commands.md b/docs/archive/1.0/api/cli/dot_commands.md index 469f1eda37d..1c9edd66c46 100644 --- a/docs/archive/1.0/api/cli/dot_commands.md +++ b/docs/archive/1.0/api/cli/dot_commands.md @@ -9,7 +9,6 @@ Dot commands are available in the DuckDB CLI client. To use one of these command ## Dot Commands -
diff --git a/docs/archive/1.0/api/cli/editing.md b/docs/archive/1.0/api/cli/editing.md index aa35847afac..b26572f88e3 100644 --- a/docs/archive/1.0/api/cli/editing.md +++ b/docs/archive/1.0/api/cli/editing.md @@ -9,7 +9,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Moving -
| Key | Action | |-----------------|------------------------------------------------------------------------| @@ -30,7 +29,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## History -
| Key | Action | |------------|--------------------------------| @@ -45,7 +43,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Changing Text -
| Key | Action | |-------------------|----------------------------------------------------------| @@ -68,7 +65,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Completing -
| Key | Action | |---------------|--------------------------------------------------------| @@ -78,7 +74,6 @@ DuckDB's CLI uses a line-editing library based on [linenoise](https://github.com ## Miscellaneous -
| Key | Action | |------------|------------------------------------------------------------------------------------| diff --git a/docs/archive/1.0/api/cli/output_formats.md b/docs/archive/1.0/api/cli/output_formats.md index d11a5223cc5..d0c2e2ed0a4 100644 --- a/docs/archive/1.0/api/cli/output_formats.md +++ b/docs/archive/1.0/api/cli/output_formats.md @@ -8,7 +8,6 @@ title: Output Formats The `.mode` [dot command]({% link docs/archive/1.0/api/cli/dot_commands.md %}) may be used to change the appearance of the tables returned in the terminal output. In addition to customizing the appearance, these modes have additional benefits. This can be useful for presenting DuckDB output elsewhere by redirecting the terminal [output to a file]({% link docs/archive/1.0/api/cli/dot_commands.md %}#output-writing-results-to-a-file). Using the `insert` mode will build a series of SQL statements that can be used to insert the data at a later point. The `markdown` mode is particularly useful for building documentation and the `latex` mode is useful for writing academic papers. -
| Mode | Description | |--------------|----------------------------------------------| diff --git a/docs/archive/1.0/api/cli/syntax_highlighting.md b/docs/archive/1.0/api/cli/syntax_highlighting.md index 1cef30903d3..9b2eaf01475 100644 --- a/docs/archive/1.0/api/cli/syntax_highlighting.md +++ b/docs/archive/1.0/api/cli/syntax_highlighting.md @@ -14,7 +14,6 @@ Syntax highlighting can also be disabled entirely using the `.highlight off` com Below is a list of components that can be configured. -
| Type | Command | Default Color | |-------------------------|-------------|-----------------| @@ -27,7 +26,6 @@ Below is a list of components that can be configured. The components can be configured using either a supported color name (e.g., `.keyword red`), or by directly providing a terminal code to use for rendering (e.g., `.keywordcode \033[31m`). Below is a list of supported color names and their corresponding terminal codes. -
| Color | Terminal Code | |---------------|---------------| diff --git a/docs/archive/1.0/configuration/pragmas.md b/docs/archive/1.0/configuration/pragmas.md index 6b2517fe779..f7c376e7744 100644 --- a/docs/archive/1.0/configuration/pragmas.md +++ b/docs/archive/1.0/configuration/pragmas.md @@ -106,7 +106,6 @@ CALL pragma_storage_info('table_name'); This call returns the following information for the given table: -
| Name | Type | Description | |----------------|-----------|-------------------------------------------------------| diff --git a/docs/archive/1.0/data/csv/auto_detection.md b/docs/archive/1.0/data/csv/auto_detection.md index 56bcb0fa60d..03d490489dc 100644 --- a/docs/archive/1.0/data/csv/auto_detection.md +++ b/docs/archive/1.0/data/csv/auto_detection.md @@ -60,7 +60,6 @@ Dialect detection works by attempting to parse the samples using the set of cons The following dialects are considered for automatic dialect detection. -
diff --git a/docs/archive/1.0/data/csv/overview.md b/docs/archive/1.0/data/csv/overview.md index d4a2c01e9ec..eec0300eca1 100644 --- a/docs/archive/1.0/data/csv/overview.md +++ b/docs/archive/1.0/data/csv/overview.md @@ -135,7 +135,6 @@ The `read_csv` automatically attempts to figure out the correct configuration of SELECT * FROM read_csv('flights.csv'); ``` -
| FlightDate | UniqueCarrier | OriginCityName | DestCityName | |------------|---------------|----------------|-----------------| @@ -153,7 +152,6 @@ CREATE TABLE ontime AS DESCRIBE ontime; ``` -
| column_name | column_type | null | key | default | extra | |----------------|-------------|------|------|---------|-------| @@ -189,7 +187,6 @@ COPY ontime FROM 'flights.csv'; SELECT * FROM ontime; ``` -
| flightdate | uniquecarrier | origincityname | destcityname | |------------|---------------|----------------|-----------------| diff --git a/docs/archive/1.0/data/csv/reading_faulty_csv_files.md b/docs/archive/1.0/data/csv/reading_faulty_csv_files.md index fd25034df6d..f4339847ad6 100644 --- a/docs/archive/1.0/data/csv/reading_faulty_csv_files.md +++ b/docs/archive/1.0/data/csv/reading_faulty_csv_files.md @@ -112,7 +112,6 @@ FROM read_csv( Outputs: -
| name | age | |-------|-----| @@ -127,7 +126,6 @@ FROM read_csv('faulty.csv', columns = {'name': 'VARCHAR', 'age': 'INTEGER'}); Outputs: -
| name | |--------------| @@ -148,7 +146,6 @@ Note that any of the errors described in our Structural Error section will be st The CSV Reject Scans Table returns the following information: -
| Column name | Description | Type | |:--|:-----|:-| @@ -170,7 +167,6 @@ The CSV Reject Scans Table returns the following information: The CSV Reject Errors Table returns the following information: -
| Column name | Description | Type | |:--|:-----|:-| @@ -187,7 +183,6 @@ The CSV Reject Errors Table returns the following information: ## Parameters -
The parameters listed below are used in the `read_csv` function to configure the CSV Rejects Table. diff --git a/docs/archive/1.0/data/multiple_files/combining_schemas.md b/docs/archive/1.0/data/multiple_files/combining_schemas.md index 13a683f9b77..0d64a1bc40b 100644 --- a/docs/archive/1.0/data/multiple_files/combining_schemas.md +++ b/docs/archive/1.0/data/multiple_files/combining_schemas.md @@ -48,7 +48,6 @@ FlightDate|UniqueCarrier|OriginCityName|DestCityName Reading the two files at the same time will produce the following result set: -
| FlightDate | UniqueCarrier | OriginCityName | DestCityName | |------------|---------------|----------------|-----------------| @@ -83,7 +82,6 @@ Reading these when unifying column names **by position** results in an error – SELECT * FROM read_csv(['flights3.csv', 'flights4.csv'], union_by_name = true); ``` -
| FlightDate | OriginCityName | DestCityName | UniqueCarrier | |------------|----------------|-----------------|---------------| diff --git a/docs/archive/1.0/data/multiple_files/overview.md b/docs/archive/1.0/data/multiple_files/overview.md index 5eeae9b36b9..568c93370d1 100644 --- a/docs/archive/1.0/data/multiple_files/overview.md +++ b/docs/archive/1.0/data/multiple_files/overview.md @@ -94,7 +94,6 @@ FROM read_parquet(['file1.parquet', 'file2.parquet', 'file3.parquet']); Any file name input to the `read_parquet` function can either be an exact filename, or use a glob syntax to read multiple files that match a pattern. -
| Wildcard | Description | |------------|-----------------------------------------------------------| @@ -137,7 +136,6 @@ SELECT * FROM read_csv(['flights1.csv', 'flights2.csv'], union_by_name = true, filename = true); ``` -
| FlightDate | OriginCityName | DestCityName | UniqueCarrier | filename | |------------|----------------|-----------------|---------------|--------------| @@ -157,7 +155,6 @@ SELECT * FROM glob('*'); ``` -
| file | |---------------| diff --git a/docs/archive/1.0/dev/release_calendar.md b/docs/archive/1.0/dev/release_calendar.md index dcd8078e9a4..097e99880a4 100644 --- a/docs/archive/1.0/dev/release_calendar.md +++ b/docs/archive/1.0/dev/release_calendar.md @@ -18,7 +18,6 @@ Patch versions only ship bugfixes, while minor versions also introduce new featu The planned dates of upcoming DuckDB releases are shown below. **Please note that these dates are tentative** and DuckDB maintainers may decide to push back release dates to ensure the stability and quality of releases. -
@@ -32,7 +31,6 @@ The planned dates of upcoming DuckDB releases are shown below. ## Past Releases -
In the following, we list DuckDB's past releases along with their codename where applicable. Between versions 0.2.2 and 0.3.3, all releases (including patch versions) received a codename. diff --git a/docs/archive/1.0/extensions/arrow.md b/docs/archive/1.0/extensions/arrow.md index 4d5e5312b76..936cd08d6d9 100644 --- a/docs/archive/1.0/extensions/arrow.md +++ b/docs/archive/1.0/extensions/arrow.md @@ -18,7 +18,6 @@ LOAD arrow; ## Functions -
| Function | Type | Description | |--|----|-------| diff --git a/docs/archive/1.0/extensions/autocomplete.md b/docs/archive/1.0/extensions/autocomplete.md index 8a2c0fd3a0d..bc2fa72b9e9 100644 --- a/docs/archive/1.0/extensions/autocomplete.md +++ b/docs/archive/1.0/extensions/autocomplete.md @@ -13,7 +13,6 @@ For the behavior of the `autocomplete` extension, see the [documentation of the ## Functions -
| Function | Description | |:----------------------------------|:-----------------------------------------------------| @@ -28,7 +27,6 @@ FROM sql_auto_complete('SEL'); Returns: -
| suggestion | suggestion_start | |-------------|------------------| diff --git a/docs/archive/1.0/extensions/core_extensions.md b/docs/archive/1.0/extensions/core_extensions.md index ebe723478ac..00fa4645961 100644 --- a/docs/archive/1.0/extensions/core_extensions.md +++ b/docs/archive/1.0/extensions/core_extensions.md @@ -37,7 +37,6 @@ title: Core Extensions Different DuckDB clients ship a different set of extensions. We summarize the main distributions in the table below. -
| Name | CLI (duckdb.org) | CLI (Homebrew) | Python | R | Java | Node.js | |------|------|------|---|---|---|---|---| diff --git a/docs/archive/1.0/extensions/full_text_search.md b/docs/archive/1.0/extensions/full_text_search.md index 1b75193316a..d6207cadaf5 100644 --- a/docs/archive/1.0/extensions/full_text_search.md +++ b/docs/archive/1.0/extensions/full_text_search.md @@ -56,7 +56,6 @@ drop_fts_index(input_table) Drops a FTS index for the specified table. -
| Name | Type | Description | |:--|:--|:-----------| diff --git a/docs/archive/1.0/extensions/httpfs/s3api.md b/docs/archive/1.0/extensions/httpfs/s3api.md index 4d1c4a89ded..859a2c74679 100644 --- a/docs/archive/1.0/extensions/httpfs/s3api.md +++ b/docs/archive/1.0/extensions/httpfs/s3api.md @@ -11,7 +11,6 @@ The `httpfs` filesystem is tested with [AWS S3](https://aws.amazon.com/s3/), [Mi The following table shows which parts of the S3 API are required for each `httpfs` feature. -
| Feature | Required S3 API features | |:---|:---| @@ -199,7 +198,6 @@ FROM read_parquet('s3://bucket/*.parquet', filename = true); could for example result in: -
| column_a | column_b | filename | |:---|:---|:---| @@ -247,7 +245,6 @@ s3://my-bucket/partitioned/part_col_a=⟨val⟩/part_col_b=⟨val⟩/data_⟨thr Some additional configuration options exist for the S3 upload, though the default values should suffice for most use cases. -
| Name | Description | |:---|:---| diff --git a/docs/archive/1.0/extensions/httpfs/s3api_legacy_authentication.md b/docs/archive/1.0/extensions/httpfs/s3api_legacy_authentication.md index cf0ce691fd5..a05196c723e 100644 --- a/docs/archive/1.0/extensions/httpfs/s3api_legacy_authentication.md +++ b/docs/archive/1.0/extensions/httpfs/s3api_legacy_authentication.md @@ -73,7 +73,6 @@ Some additional configuration options exist for the S3 upload, though the defaul Additionally, most of the configuration options can be set via environment variables: -
| DuckDB setting | Environment variable | Note | |:-----------------------|:------------------------|:-----------------------------------------| diff --git a/docs/archive/1.0/extensions/json.md b/docs/archive/1.0/extensions/json.md index d7432dcd4d1..a8d64dad650 100644 --- a/docs/archive/1.0/extensions/json.md +++ b/docs/archive/1.0/extensions/json.md @@ -384,7 +384,6 @@ COPY test FROM 'my.json' (AUTO_DETECT true); The following functions are used to create JSON. -
| Function | Description | |:--|:----| @@ -820,7 +819,6 @@ true There are three JSON aggregate functions. -
| Function | Description | |:---|:----| @@ -871,7 +869,6 @@ SELECT json_group_structure(j) FROM example2; In many cases, it is inefficient to extract values from JSON one-by-one. Instead, we can “extract” all values at once, transforming JSON to the nested types `LIST` and `STRUCT`. -
| Function | Description | |:---|:---| diff --git a/docs/archive/1.0/extensions/overview.md b/docs/archive/1.0/extensions/overview.md index db652183011..a6529c7c7f7 100644 --- a/docs/archive/1.0/extensions/overview.md +++ b/docs/archive/1.0/extensions/overview.md @@ -23,7 +23,6 @@ SELECT extension_name, installed, description FROM duckdb_extensions(); ``` -
| extension_name | installed | description | |-------------------|-----------|--------------------------------------------------------------| diff --git a/docs/archive/1.0/extensions/spatial.md b/docs/archive/1.0/extensions/spatial.md index b69ab92032d..5c8aeaf89cb 100644 --- a/docs/archive/1.0/extensions/spatial.md +++ b/docs/archive/1.0/extensions/spatial.md @@ -138,7 +138,6 @@ Compute relationships and spatial predicates between geometries. ## Spatial Aggregate Functions -
| Aggregate functions | Implemented with | |-------------------------------------------|------------------| diff --git a/docs/archive/1.0/extensions/tpch.md b/docs/archive/1.0/extensions/tpch.md index 20812df4c2b..aba7969060e 100644 --- a/docs/archive/1.0/extensions/tpch.md +++ b/docs/archive/1.0/extensions/tpch.md @@ -80,7 +80,6 @@ This function returns a table with columns `query_nr`, `scale_factor`, and `answ The data generator function `dbgen` has the following parameters: -
| Name | Type | Description | |--|--|------------| diff --git a/docs/archive/1.0/extensions/versioning_of_extensions.md b/docs/archive/1.0/extensions/versioning_of_extensions.md index d99d38d48f5..f5f0a2693d9 100644 --- a/docs/archive/1.0/extensions/versioning_of_extensions.md +++ b/docs/archive/1.0/extensions/versioning_of_extensions.md @@ -17,7 +17,6 @@ FROM duckdb_extensions() WHERE extension_name='parquet'; ``` -
| extension_name | extension_version | install_mode | |:------------------|:------------------|:---------------------| @@ -33,7 +32,6 @@ FROM duckdb_extensions() WHERE extension_name = 'azure'; ``` -
| extension_name | extension_version | install_mode | |:---------------|:------------------|:---------------| @@ -50,7 +48,6 @@ give the user information on which extensions were updated to/from which version UPDATE EXTENSIONS; ``` -
| extension_name | repository | update_result | previous_version | current_version | |:---------------|:-------------|:----------------------|:-----------------|:----------------| @@ -68,7 +65,6 @@ The update statement can also be provided with a list of specific extensions to UPDATE EXTENSIONS (httpfs, azure); ``` -
| extension_name | repository | update_result | previous_version | current_version | |:---------------|:-------------|:----------------------|:-----------------|:----------------| diff --git a/docs/archive/1.0/extensions/working_with_extensions.md b/docs/archive/1.0/extensions/working_with_extensions.md index 2cd20208263..9fe48212909 100644 --- a/docs/archive/1.0/extensions/working_with_extensions.md +++ b/docs/archive/1.0/extensions/working_with_extensions.md @@ -10,7 +10,6 @@ For platforms where packages for certain extensions are not available, users can All official extensions are distributed for the following platforms. -
| Platform name | Operating system | Architecture | CPU types | Used by | |--------------------|------------------|-----------------|---------------------------------|------------------------------| @@ -89,7 +88,6 @@ SET custom_extension_repository = 'http://nightly-extensions.duckdb.org'; While any url or local path can be used as a repository, currently DuckDB contains the following predefined repositories: -
| Alias | Url | Description | |:----------------------|:---------------------------------------|:---------------------------------------------------------------------------------------| diff --git a/docs/archive/1.0/guides/database_integration/mysql.md b/docs/archive/1.0/guides/database_integration/mysql.md index dd89f5dd780..dd1bb27a4fc 100644 --- a/docs/archive/1.0/guides/database_integration/mysql.md +++ b/docs/archive/1.0/guides/database_integration/mysql.md @@ -32,7 +32,6 @@ USE mysql_db; The string used by `ATTACH` is a PostgreSQL-style connection string (_not_ a MySQL connection string!). It is a list of connection arguments provided in `{key}={value}` format. Below is a list of valid arguments. Any options not provided are replaced by their default values. -
| Setting | Default | |------------|--------------| diff --git a/docs/archive/1.0/guides/database_integration/postgres.md b/docs/archive/1.0/guides/database_integration/postgres.md index a91bb00c243..0c54718085d 100644 --- a/docs/archive/1.0/guides/database_integration/postgres.md +++ b/docs/archive/1.0/guides/database_integration/postgres.md @@ -32,7 +32,6 @@ SELECT * FROM postgres_scan('host=localhost port=5432 dbname=mydb', 'public', 'm The first parameter to the `postgres_scan` function is the [PostgreSQL connection string](https://www.postgresql.org/docs/current/libpq-connect.html#LIBPQ-CONNSTRING), a list of connection arguments provided in `{key}={value}` format. Below is a list of valid arguments. -
| Name | Description | Default | |------------|--------------------------------------|----------------| diff --git a/docs/archive/1.0/guides/meta/describe.md b/docs/archive/1.0/guides/meta/describe.md index aa58358a1e3..3746ed7bf91 100644 --- a/docs/archive/1.0/guides/meta/describe.md +++ b/docs/archive/1.0/guides/meta/describe.md @@ -13,7 +13,6 @@ DESCRIBE tbl; SHOW tbl; -- equivalent to DESCRIBE tbl; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|------|---------|-------| @@ -28,7 +27,6 @@ In order to view the schema of the result of a query, prepend `DESCRIBE` to a qu DESCRIBE SELECT * FROM tbl; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|------|---------|-------| diff --git a/docs/archive/1.0/guides/meta/list_tables.md b/docs/archive/1.0/guides/meta/list_tables.md index f57bf63ae15..6a75dc92822 100644 --- a/docs/archive/1.0/guides/meta/list_tables.md +++ b/docs/archive/1.0/guides/meta/list_tables.md @@ -10,7 +10,6 @@ CREATE TABLE tbl (i INTEGER); SHOW TABLES; ``` -
| name | |------| @@ -25,7 +24,6 @@ CREATE TABLE s1.tbl (v VARCHAR); SHOW ALL TABLES; ``` -
| database | schema | table_name | column_names | column_types | temporary | |----------|--------|------------|--------------|--------------|-----------| diff --git a/docs/archive/1.0/guides/network_cloud_storage/s3_express_one.md b/docs/archive/1.0/guides/network_cloud_storage/s3_express_one.md index 8cee09fe754..5a23ea1b09d 100644 --- a/docs/archive/1.0/guides/network_cloud_storage/s3_express_one.md +++ b/docs/archive/1.0/guides/network_cloud_storage/s3_express_one.md @@ -70,7 +70,6 @@ FROM 's3://express-bucket-name--use1-az5--x-s3/my-file.parquet'; We ran two experiments on a `c7gd.12xlarge` instance using the [LDBC SF300 Comments `creationDate` Parquet file](https://blobs.duckdb.org/data/ldbc-sf300-comments-creationDate.parquet) file (also used in the [microbenchmarks of the performance guide]({% link docs/archive/1.0/guides/performance/benchmarks.md %}#data-sets)). -
| Experiment | File size | Runtime | |:-----|--:|--:| diff --git a/docs/archive/1.0/guides/performance/file_formats.md b/docs/archive/1.0/guides/performance/file_formats.md index 9130b26bc0c..6e7e5fdbf80 100644 --- a/docs/archive/1.0/guides/performance/file_formats.md +++ b/docs/archive/1.0/guides/performance/file_formats.md @@ -40,7 +40,6 @@ DuckDB works best on Parquet files with row groups of 100K-1M rows each. The rea We run a simple aggregation query over Parquet files using different row group sizes, selected between 960 and 1,966,080. The results are as follows. -
| Row group size | Execution time | |---------------:|---------------:| @@ -77,7 +76,6 @@ For tips on reading and writing Parquet files, see the [Parquet Tips page]({% li CSV files are often distributed in compressed format such as GZIP archives (`.csv.gz`). DuckDB can decompress these files on the fly. In fact, this is typically faster than decompressing the files first and loading them due to reduced IO. -
| Schema | Load Time | |---|--:| diff --git a/docs/archive/1.0/guides/performance/indexing.md b/docs/archive/1.0/guides/performance/indexing.md index 98bdd9979be..8123afffeda 100644 --- a/docs/archive/1.0/guides/performance/indexing.md +++ b/docs/archive/1.0/guides/performance/indexing.md @@ -17,7 +17,6 @@ The more ordered the data within a column, the more useful the zonemap indexes w For an example, let’s repeat the [microbenchmark for timestamps]({% link docs/archive/1.0/guides/performance/schema.md %}#microbenchmark-using-timestamps) with a timestamp column that sorted using an ascending order vs. an unordered one. -
| Column type | Ordered | Storage size | Query time | |---|---|--:|--:| diff --git a/docs/archive/1.0/guides/performance/schema.md b/docs/archive/1.0/guides/performance/schema.md index ed7eef84432..ebe6762b051 100644 --- a/docs/archive/1.0/guides/performance/schema.md +++ b/docs/archive/1.0/guides/performance/schema.md @@ -31,7 +31,6 @@ SELECT avg(CAST(creationDate[9:10] AS INTEGER)) FROM Comment; The results of the microbenchmark are as follows: -
| Column type | Storage size | Query time | | ----------- | -----------: | ---------: | @@ -55,7 +54,6 @@ In the second experiment, we define all columns with the `VARCHAR` type. While the results of the queries are the same for all both experiments, their runtime vary significantly. The results below show that joining on `BIGINT` columns is approx. 1.8× faster than performing the same join on `VARCHAR`-typed columns encoding the same value. -
| Join column payload type | Join column schema type | Example value | Query time | | ------------------------ | ----------------------- | ---------------------------------------- | ---------: | @@ -74,7 +72,6 @@ DuckDB allows defining [constraints]({% link docs/archive/1.0/sql/constraints.md We illustrate the effect of using primary keys with the [LDBC Comment table at scale factor 300](https://blobs.duckdb.org/data/ldbc-sf300-comments.tar.zst). This table has approx. 554 million entries. We first create the schema without a primary key, then load the data. In the second experiment, we create the schema with a primary key, then load the data. In both cases, we take the data from `.csv.gz` files, and measure the time required to perform the loading. -
| Operation | Execution time | | ------------------------ | -------------: | diff --git a/docs/archive/1.0/guides/sql_features/asof_join.md b/docs/archive/1.0/guides/sql_features/asof_join.md index 93b3d402cd4..f8a42906188 100644 --- a/docs/archive/1.0/guides/sql_features/asof_join.md +++ b/docs/archive/1.0/guides/sql_features/asof_join.md @@ -24,7 +24,6 @@ which can be cumbersome and slow to implement in standard SQL. Let's start with a concrete example. Suppose we have a table of stock [`prices`](/data/prices.csv) with timestamps: -
| ticker | when | price | | :----- | :--- | ----: | @@ -40,7 +39,6 @@ Suppose we have a table of stock [`prices`](/data/prices.csv) with timestamps: We have another table containing portfolio [`holdings`](/data/holdings.csv) at various points in time: -
| ticker | when | shares | | :----- | :--- | -----: | @@ -76,7 +74,6 @@ ASOF JOIN prices p This attaches the value of the holding at that time to each row: -
| ticker | when | value | | :----- | :--- | ----: | @@ -107,7 +104,6 @@ ORDER BY ALL; As you might expect, this will produce `NULL` prices and values instead of dropping left side rows when there is no ticker or the time is before the prices begin. -
| ticker | when | value | | :----- | :--- | ----: | diff --git a/docs/archive/1.0/internals/storage.md b/docs/archive/1.0/internals/storage.md index 8d38950de53..74a8d8e7a98 100644 --- a/docs/archive/1.0/internals/storage.md +++ b/docs/archive/1.0/internals/storage.md @@ -67,7 +67,6 @@ with open('test/sql/storage_version/storage_version.db', 'rb') as fh: For changes in each given release, check out the [change log](https://github.com/duckdb/duckdb/releases) on GitHub. To see the commits that changed each storage version, see the [commit log](https://github.com/duckdb/duckdb/commits/main/src/storage/storage_info.cpp). -
| Storage version | DuckDB version(s) | |----------------:|---------------------------------| diff --git a/docs/archive/1.0/sql/data_types/bitstring.md b/docs/archive/1.0/sql/data_types/bitstring.md index 522eed99f20..72c9c48759d 100644 --- a/docs/archive/1.0/sql/data_types/bitstring.md +++ b/docs/archive/1.0/sql/data_types/bitstring.md @@ -4,7 +4,6 @@ layout: docu title: Bitstring Type --- -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/1.0/sql/data_types/blob.md b/docs/archive/1.0/sql/data_types/blob.md index b07c376702c..2ffc4d37890 100644 --- a/docs/archive/1.0/sql/data_types/blob.md +++ b/docs/archive/1.0/sql/data_types/blob.md @@ -5,7 +5,6 @@ layout: docu title: Blob Type --- -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/1.0/sql/data_types/boolean.md b/docs/archive/1.0/sql/data_types/boolean.md index 38ee01ed206..69ea829756b 100644 --- a/docs/archive/1.0/sql/data_types/boolean.md +++ b/docs/archive/1.0/sql/data_types/boolean.md @@ -4,7 +4,6 @@ layout: docu title: Boolean Type --- -
| Name | Aliases | Description | |:---|:---|:---| @@ -45,7 +44,6 @@ The `AND`/`OR` conjunctions can be used to combine boolean values. Below is the truth table for the `AND` conjunction (i.e., `x AND y`). -
| X | X AND true | X AND false | X AND NULL | |-------|-------|-------|-------| @@ -55,7 +53,6 @@ Below is the truth table for the `AND` conjunction (i.e., `x AND y`). Below is the truth table for the `OR` conjunction (i.e., `x OR y`). -
| X | X OR true | X OR false | X OR NULL | |-------|------|-------|------| diff --git a/docs/archive/1.0/sql/data_types/date.md b/docs/archive/1.0/sql/data_types/date.md index 6678d3db507..35997d580c9 100644 --- a/docs/archive/1.0/sql/data_types/date.md +++ b/docs/archive/1.0/sql/data_types/date.md @@ -4,7 +4,6 @@ layout: docu title: Date Types --- -
| Name | Aliases | Description | |:-------|:--------|:--------------------------------| @@ -20,7 +19,6 @@ SELECT DATE '1992-09-20'; There are also three special date values that can be used on input: -
| Input string | Description | |:-------------|:----------------------------------| diff --git a/docs/archive/1.0/sql/data_types/enum.md b/docs/archive/1.0/sql/data_types/enum.md index 37ef046c5f1..766c090928e 100644 --- a/docs/archive/1.0/sql/data_types/enum.md +++ b/docs/archive/1.0/sql/data_types/enum.md @@ -5,7 +5,6 @@ layout: docu title: Enum Data Type --- -
| Name | Description | |:--|:-----| @@ -76,7 +75,6 @@ Show the available values in the `birds` enum using the `enum_range` function: SELECT enum_range(NULL::birds) AS my_enum_range; ``` -
| my_enum_range | |-----------------| diff --git a/docs/archive/1.0/sql/data_types/interval.md b/docs/archive/1.0/sql/data_types/interval.md index 8afbbe8f59f..74c19f29ebf 100644 --- a/docs/archive/1.0/sql/data_types/interval.md +++ b/docs/archive/1.0/sql/data_types/interval.md @@ -7,7 +7,6 @@ title: Interval Type `INTERVAL`s represent periods of time that can be added to or subtracted from `DATE`, `TIMESTAMP`, `TIMESTAMPTZ`, or `TIME` values. -
| Name | Description | |:---|:---| diff --git a/docs/archive/1.0/sql/data_types/numeric.md b/docs/archive/1.0/sql/data_types/numeric.md index 567e5bd271e..17373b3c533 100644 --- a/docs/archive/1.0/sql/data_types/numeric.md +++ b/docs/archive/1.0/sql/data_types/numeric.md @@ -30,7 +30,6 @@ The data type `DECIMAL(WIDTH, SCALE)` (also available under the alias `NUMERIC(W Internally, decimals are represented as integers depending on their specified width. -
| Width | Internal | Size (bytes) | |:---|:---|---:| @@ -45,7 +44,6 @@ Performance can be impacted by using too large decimals when not required. In pa The data types `FLOAT` and `DOUBLE` precision are variable-precision numeric types. In practice, these types are usually implementations of IEEE Standard 754 for Binary Floating-Point Arithmetic (single and double precision, respectively), to the extent that the underlying processor, operating system, and compiler support it. -
| Name | Aliases | Description | |:--|:--|:--------| diff --git a/docs/archive/1.0/sql/data_types/struct.md b/docs/archive/1.0/sql/data_types/struct.md index 5b3b208acce..88b799a7c3c 100644 --- a/docs/archive/1.0/sql/data_types/struct.md +++ b/docs/archive/1.0/sql/data_types/struct.md @@ -111,7 +111,6 @@ SELECT a.* FROM (SELECT {'x': 1, 'y': 2, 'z': 3} AS a); ``` -
| x | y | z | |:---|:---|:---| diff --git a/docs/archive/1.0/sql/data_types/text.md b/docs/archive/1.0/sql/data_types/text.md index 043453e0107..94d43a2ea6e 100644 --- a/docs/archive/1.0/sql/data_types/text.md +++ b/docs/archive/1.0/sql/data_types/text.md @@ -7,7 +7,6 @@ title: Text Types In DuckDB, strings can be stored in the `VARCHAR` field. The field allows storage of Unicode characters. Internally, the data is encoded as UTF-8. -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/archive/1.0/sql/data_types/time.md b/docs/archive/1.0/sql/data_types/time.md index db4e22de76c..978ea2e4cbd 100644 --- a/docs/archive/1.0/sql/data_types/time.md +++ b/docs/archive/1.0/sql/data_types/time.md @@ -6,7 +6,6 @@ title: Time Types The `TIME` and `TIMETZ` types specify the hour, minute, second, microsecond of a day. -
| Name | Aliases | Description | | :------- | :----------------------- | :------------------------------ | diff --git a/docs/archive/1.0/sql/data_types/timestamp.md b/docs/archive/1.0/sql/data_types/timestamp.md index 4ffb9f93799..ed77a96967f 100644 --- a/docs/archive/1.0/sql/data_types/timestamp.md +++ b/docs/archive/1.0/sql/data_types/timestamp.md @@ -82,7 +82,6 @@ SELECT TIMESTAMP WITH TIME ZONE '1992-09-20 11:30:00.123456789'; There are also three special date values that can be used on input: -
| Input string | Valid types | Description | |:-------------|:---------------------------|:-----------------------------------------------| @@ -97,7 +96,6 @@ but `epoch` is simply a notational shorthand that will be converted to the time SELECT '-infinity'::TIMESTAMP, 'epoch'::TIMESTAMP, 'infinity'::TIMESTAMP; ``` -
| Negative | Epoch | Positive | |:----------|:--------------------|:---------| diff --git a/docs/archive/1.0/sql/data_types/timezones.md b/docs/archive/1.0/sql/data_types/timezones.md index e81a9a738c6..3c0b3409572 100644 --- a/docs/archive/1.0/sql/data_types/timezones.md +++ b/docs/archive/1.0/sql/data_types/timezones.md @@ -12,7 +12,6 @@ FROM pg_timezone_names() ORDER BY name; ``` -
| name | abbrev | |----------------------------------|----------------------------------| diff --git a/docs/archive/1.0/sql/data_types/union.md b/docs/archive/1.0/sql/data_types/union.md index 1302d14e60e..d2b6b500507 100644 --- a/docs/archive/1.0/sql/data_types/union.md +++ b/docs/archive/1.0/sql/data_types/union.md @@ -95,7 +95,6 @@ So how do we disambiguate if we want to create a `UNION` with multiple members o `UNION` types can be cast between each other if the source type is a “subset” of the target type. In other words, all the tags in the source `UNION` must be present in the target `UNION`, and all the types of the matching tags must be implicitly castable between source and target. In essence, this means that `UNION` types are covariant with respect to their members. -
| Ok | Source | Target | Comments | |----|------------------------|------------------------|----------------------------------------| diff --git a/docs/archive/1.0/sql/expressions/comparison_operators.md b/docs/archive/1.0/sql/expressions/comparison_operators.md index 593a8b1735d..f376fd02dc7 100644 --- a/docs/archive/1.0/sql/expressions/comparison_operators.md +++ b/docs/archive/1.0/sql/expressions/comparison_operators.md @@ -11,7 +11,6 @@ title: Comparisons The table below shows the standard comparison operators. Whenever either of the input arguments is `NULL`, the output of the comparison is `NULL`. -
| Operator | Description | Example | Result | |:---|:---|:---|:---| @@ -25,7 +24,6 @@ Whenever either of the input arguments is `NULL`, the output of the comparison i The table below shows the standard distinction operators. These operators treat `NULL` values as equal. -
| Operator | Description | Example | Result | |:---|:---|:---|:-| @@ -40,7 +38,6 @@ Besides the standard comparison operators there are also the `BETWEEN` and `IS ( Note that `BETWEEN` and `NOT BETWEEN` are only equivalent to the examples below in the cases where both `a`, `x` and `y` are of the same type, as `BETWEEN` will cast all of its inputs to the same type. -
| Predicate | Description | |:---|:---| diff --git a/docs/archive/1.0/sql/expressions/logical_operators.md b/docs/archive/1.0/sql/expressions/logical_operators.md index 57a6a4f1f88..5249033ac8b 100644 --- a/docs/archive/1.0/sql/expressions/logical_operators.md +++ b/docs/archive/1.0/sql/expressions/logical_operators.md @@ -10,7 +10,6 @@ The following logical operators are available: `AND`, `OR` and `NOT`. SQL uses a ### Binary Operators: `AND` and `OR` -
| `a` | `b` | `a AND b` | `a OR b` | |:---|:---|:---|:---| @@ -23,7 +22,6 @@ The following logical operators are available: `AND`, `OR` and `NOT`. SQL uses a ### Unary Operator: `NOT` -
| `a` | `NOT a` | |:---|:---| diff --git a/docs/archive/1.0/sql/expressions/star.md b/docs/archive/1.0/sql/expressions/star.md index e1efc30ecac..573d878cda5 100644 --- a/docs/archive/1.0/sql/expressions/star.md +++ b/docs/archive/1.0/sql/expressions/star.md @@ -103,7 +103,6 @@ INSERT INTO numbers VALUES (1, 10), (2, 20), (3, NULL); SELECT min(COLUMNS(*)), count(COLUMNS(*)) FROM numbers; ``` -
| id | number | id | number | |---:|-------:|---:|-------:| @@ -118,7 +117,6 @@ SELECT FROM numbers; ``` -
| id | min(number := (number + id)) | id | |---:|-----------------------------:|---:| @@ -130,7 +128,6 @@ FROM numbers; SELECT COLUMNS(*) + COLUMNS(*) FROM numbers; ``` -
| id | number | |---:|-------:| @@ -152,7 +149,6 @@ FROM ( WHERE COLUMNS(*) > 1; -- equivalent to: x > 1 AND y > 1 AND z > 1 ``` -
| x | y | z | |--:|--:|--:| @@ -166,7 +162,6 @@ WHERE COLUMNS(*) > 1; -- equivalent to: x > 1 AND y > 1 AND z > 1 SELECT COLUMNS('(id|numbers?)') FROM numbers; ``` -
| id | number | |---:|-------:| @@ -180,7 +175,6 @@ The matches of capture groups can be used to rename columns selected by a regula SELECT COLUMNS('(\w{2}).*') AS '\1' FROM numbers; ``` -
| id | nu | |---:|-----:| @@ -198,7 +192,6 @@ The capture groups are one-indexed; `\0` is the original column name. SELECT COLUMNS(c -> c LIKE '%num%') FROM numbers; ``` -
| number | |--------| @@ -218,7 +211,6 @@ For example: SELECT st.* FROM (SELECT {'x': 1, 'y': 2, 'z': 3} AS st); ``` -
| x | y | z | |--:|--:|--:| diff --git a/docs/archive/1.0/sql/expressions/subqueries.md b/docs/archive/1.0/sql/expressions/subqueries.md index 0aec6267e17..f73dfdf2bc7 100644 --- a/docs/archive/1.0/sql/expressions/subqueries.md +++ b/docs/archive/1.0/sql/expressions/subqueries.md @@ -16,7 +16,6 @@ Consider the following table: ### Grades -
| grade | course | |---:|:---| diff --git a/docs/archive/1.0/sql/functions/bitstring.md b/docs/archive/1.0/sql/functions/bitstring.md index 9625766865a..9dad70eaf2a 100644 --- a/docs/archive/1.0/sql/functions/bitstring.md +++ b/docs/archive/1.0/sql/functions/bitstring.md @@ -14,7 +14,6 @@ Bitstrings must be of equal length when performing the bitwise operands AND, OR The table below shows the available mathematical operators for `BIT` type. -
diff --git a/docs/archive/1.0/sql/functions/char.md b/docs/archive/1.0/sql/functions/char.md index 8d0aac5bd1d..11acb99abd9 100644 --- a/docs/archive/1.0/sql/functions/char.md +++ b/docs/archive/1.0/sql/functions/char.md @@ -855,7 +855,6 @@ SELECT format('I''d rather be {1} than {0}.', 'right', 'happy'); -- I'd rather b #### Format Specifiers -
| Specifier | Description | Example | |:-|:------|:---| @@ -966,7 +965,6 @@ I'd rather be happy than right. #### Format Specifiers -
| Specifier | Description | Example | |:-|:------|:---| diff --git a/docs/archive/1.0/sql/functions/dateformat.md b/docs/archive/1.0/sql/functions/dateformat.md index d54987e5cbc..be5ad9f9115 100644 --- a/docs/archive/1.0/sql/functions/dateformat.md +++ b/docs/archive/1.0/sql/functions/dateformat.md @@ -85,7 +85,6 @@ FROM read_csv('test.csv', dateformat = '%m/%d/%Y'); Below is a full list of all available format specifiers. -
| Specifier | Description | Example | |:-|:------|:---| diff --git a/docs/archive/1.0/sql/functions/numeric.md b/docs/archive/1.0/sql/functions/numeric.md index 10fabd5e1e3..13a1219122e 100644 --- a/docs/archive/1.0/sql/functions/numeric.md +++ b/docs/archive/1.0/sql/functions/numeric.md @@ -11,7 +11,6 @@ title: Numeric Functions The table below shows the available mathematical operators for [numeric types]({% link docs/archive/1.0/sql/data_types/numeric.md %}). -
diff --git a/docs/archive/1.0/sql/functions/pattern_matching.md b/docs/archive/1.0/sql/functions/pattern_matching.md index 5fe4d6c6183..08f59610e58 100644 --- a/docs/archive/1.0/sql/functions/pattern_matching.md +++ b/docs/archive/1.0/sql/functions/pattern_matching.md @@ -144,7 +144,6 @@ SELECT NOT 'best.txt' GLOB '*.txt'; -- false Three tildes (`~~~`) may also be used in place of the `GLOB` keyword. -
| GLOB-style | Symbolic-style | |:---|:---| diff --git a/docs/archive/1.0/sql/functions/regular_expressions.md b/docs/archive/1.0/sql/functions/regular_expressions.md index 16949f97812..95544fbca15 100644 --- a/docs/archive/1.0/sql/functions/regular_expressions.md +++ b/docs/archive/1.0/sql/functions/regular_expressions.md @@ -113,7 +113,6 @@ SELECT regexp_matches('abc', 'A', 'i'); -- true The regex functions support the following `options`. -
| Option | Description | |:---|:---| @@ -138,7 +137,6 @@ SELECT regexp_matches(E'hello\nworld', 'hello.world', 's'); -- true The `regexp_matches` operator will be optimized to the `LIKE` operator when possible. To achieve best performance, the `'c'` option (case-sensitive matching) should be passed if applicable. Note that by default the [`RE2` library](#regular-expression-syntax) doesn't match the `.` character to newline. -
| Original | Optimized equivalent | |:---|:---| diff --git a/docs/archive/1.0/sql/functions/time.md b/docs/archive/1.0/sql/functions/time.md index c71d671ae12..d896404f4fe 100644 --- a/docs/archive/1.0/sql/functions/time.md +++ b/docs/archive/1.0/sql/functions/time.md @@ -11,7 +11,6 @@ This section describes functions and operators for examining and manipulating [` The table below shows the available mathematical operators for `TIME` types. -
| Operator | Description | Example | Result | |:-|:---|:----|:--| diff --git a/docs/archive/1.0/sql/functions/utility.md b/docs/archive/1.0/sql/functions/utility.md index 73324abd108..69d60b53713 100644 --- a/docs/archive/1.0/sql/functions/utility.md +++ b/docs/archive/1.0/sql/functions/utility.md @@ -285,7 +285,6 @@ The functions below are difficult to categorize into specific function types and A table function is used in place of a table in a `FROM` clause. -
| Name | Description | |:--|:-------| diff --git a/docs/archive/1.0/sql/functions/window_functions.md b/docs/archive/1.0/sql/functions/window_functions.md index 408539f2735..89ef643addb 100644 --- a/docs/archive/1.0/sql/functions/window_functions.md +++ b/docs/archive/1.0/sql/functions/window_functions.md @@ -218,7 +218,6 @@ CREATE TABLE "Generation History" AS After partitioning by plant and ordering by date, it will have this layout: -
| Plant | Date | MWh | |:---|:---|---:| @@ -264,7 +263,6 @@ ORDER BY 1, 2; The result will be the following: -
| Plant | Date | Row | |:---|:---|---:| @@ -336,7 +334,6 @@ and uses a `RANGE` frame of three days on either side of each day for the `avg` (to handle any missing days). This is the result: -
| Plant | Date | MWh 7-day Moving Average | |:---|:---|---:| diff --git a/docs/archive/1.0/sql/meta/duckdb_table_functions.md b/docs/archive/1.0/sql/meta/duckdb_table_functions.md index 2711db23337..7906b2b94a5 100644 --- a/docs/archive/1.0/sql/meta/duckdb_table_functions.md +++ b/docs/archive/1.0/sql/meta/duckdb_table_functions.md @@ -170,7 +170,6 @@ Note that `duckdb_indexes` only provides metadata about secondary indexes – i. The `duckdb_keywords()` function provides metadata about DuckDB's keywords and reserved words. -
| Column | Description | Type | |:-|:---|:-| @@ -192,7 +191,6 @@ The `duckdb_memory()` function provides metadata about DuckDB's buffer manager. The `duckdb_optimizers()` function provides metadata about the optimization rules (e.g., `expression_rewriter`, `filter_pushdown`) available in the DuckDB instance. These can be selectively turned off using [`PRAGMA disabled_optimizers`]({% link docs/archive/1.0/configuration/pragmas.md %}#selectively-disabling-optimizers). -
| Column | Description | Type | |:-|:---|:-| @@ -258,7 +256,6 @@ Attributes like `temporary`, `start_value` etc. correspond to the various option The `duckdb_settings()` function provides metadata about the settings available in the DuckDB instance. -
| Column | Description | Type | |:-|:---|:-| @@ -296,7 +293,6 @@ The [`information_schema.tables`]({% link docs/archive/1.0/sql/meta/information_ The `duckdb_temporary_files()` function provides metadata about the temporary files DuckDB has written to disk, to offload data from memory. This function mostly exists for debugging and testing purposes. -
| Column | Description | Type | |:-|:---|:-| diff --git a/docs/archive/1.0/sql/query_syntax/orderby.md b/docs/archive/1.0/sql/query_syntax/orderby.md index 9ac47c41e0a..6f2df2a9879 100644 --- a/docs/archive/1.0/sql/query_syntax/orderby.md +++ b/docs/archive/1.0/sql/query_syntax/orderby.md @@ -103,7 +103,6 @@ FROM addresses ORDER BY ALL; ``` -
| address | city | zip | |------------------------|-----------|------------| @@ -120,7 +119,6 @@ FROM addresses ORDER BY ALL DESC; ``` -
| address | city | zip | |------------------------|-----------|------------| diff --git a/docs/archive/1.0/sql/query_syntax/qualify.md b/docs/archive/1.0/sql/query_syntax/qualify.md index 2ec527d08bc..6d92ad12814 100644 --- a/docs/archive/1.0/sql/query_syntax/qualify.md +++ b/docs/archive/1.0/sql/query_syntax/qualify.md @@ -88,7 +88,6 @@ WHERE function_rank < 3; ``` -
| schema_name | function_name | function_rank | |:---|:---|:---| diff --git a/docs/archive/1.0/sql/statements/create_index.md b/docs/archive/1.0/sql/statements/create_index.md index d9a72468bdc..ef3e062ac13 100644 --- a/docs/archive/1.0/sql/statements/create_index.md +++ b/docs/archive/1.0/sql/statements/create_index.md @@ -38,7 +38,6 @@ CREATE INDEX i_index ON integers ((j + k)); ### Parameters -
| Name | Description | |:-|:-----| @@ -68,7 +67,6 @@ DROP INDEX title_idx; ### Parameters -
| Name | Description | |:---|:---| diff --git a/docs/archive/1.0/sql/statements/insert.md b/docs/archive/1.0/sql/statements/insert.md index 3971bb754fc..ef330f5717b 100644 --- a/docs/archive/1.0/sql/statements/insert.md +++ b/docs/archive/1.0/sql/statements/insert.md @@ -388,7 +388,6 @@ INSERT INTO t1 RETURNING *; ``` -
| i | |---:| @@ -403,7 +402,6 @@ INSERT INTO t2 RETURNING *, i * j AS i_times_j; ``` -
| i | j | i_times_j | |--:|--:|----------:| @@ -421,7 +419,6 @@ INSERT INTO t3 RETURNING *; ``` -
| i | j | |--:|---:| diff --git a/docs/archive/1.0/sql/statements/pivot.md b/docs/archive/1.0/sql/statements/pivot.md index ef932fb061b..1149d82743e 100644 --- a/docs/archive/1.0/sql/statements/pivot.md +++ b/docs/archive/1.0/sql/statements/pivot.md @@ -53,7 +53,6 @@ INSERT INTO Cities VALUES FROM Cities; ``` -
| Country | Name | Year | Population | |---------|---------------|-----:|-----------:| @@ -83,7 +82,6 @@ ON Year USING sum(Population); ``` -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -117,7 +115,6 @@ USING sum(Population) GROUP BY Country; ``` -
| Country | 2000 | 2010 | 2020 | |---------|-----:|-----:|-----:| @@ -136,7 +133,6 @@ USING sum(Population) GROUP BY Country; ``` -
| Country | 2000 | 2010 | |---------|-----:|-----:| @@ -161,7 +157,6 @@ ON Country, Name USING sum(Population); ``` -
| Year | NL_Amsterdam | NL_New York City | NL_Seattle | US_Amsterdam | US_New York City | US_Seattle | |-----:|-------------:|------------------|------------|--------------|-----------------:|-----------:| @@ -180,7 +175,6 @@ In this case, concatenating with an underscore is used to imitate the naming con PIVOT Cities ON Country || '_' || Name USING sum(Population); ``` -
| Year | NL_Amsterdam | US_New York City | US_Seattle | |-----:|-------------:|-----------------:|-----------:| @@ -203,7 +197,6 @@ USING sum(Population) AS total, max(Population) AS max GROUP BY Country; ``` -
| Country | 2000_total | 2000_max | 2010_total | 2010_max | 2020_total | 2020_max | |---------|-----------:|---------:|-----------:|---------:|-----------:|---------:| @@ -222,7 +215,6 @@ USING sum(Population) GROUP BY Country, Name; ``` -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -272,7 +264,6 @@ JOIN (PIVOT Cities ON Name USING sum(Population) GROUP BY Country) name_pivot USING (Country); ``` -
| Country | 2000 | 2010 | 2020 | Amsterdam | New York City | Seattle | |---------|-----:|-----:|-----:|----------:|--------------:|--------:| @@ -329,7 +320,6 @@ GROUP BY ALL; This produces the result: -
| Country | Name | list("YEAR") | list(population_sum) | |---------|---------------|--------------------|----------------------| @@ -339,7 +329,6 @@ This produces the result: The `PhysicalPivot` operator converts those lists into column names and values to return this result: -
| Country | Name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -387,7 +376,6 @@ PIVOT ( ); ``` -
| Country | 2000 | 2010 | 2020 | |---------|-----:|-----:|-----:| diff --git a/docs/archive/1.0/sql/statements/unpivot.md b/docs/archive/1.0/sql/statements/unpivot.md index f8d5256fae2..da106985540 100644 --- a/docs/archive/1.0/sql/statements/unpivot.md +++ b/docs/archive/1.0/sql/statements/unpivot.md @@ -46,7 +46,6 @@ INSERT INTO monthly_sales VALUES FROM monthly_sales; ``` -
| empid | dept | Jan | Feb | Mar | Apr | May | Jun | |------:|-------------|----:|----:|----:|----:|----:|----:| @@ -72,7 +71,6 @@ INTO VALUE sales; ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -111,7 +109,6 @@ INTO VALUE sales; ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -153,7 +150,6 @@ UNPIVOT monthly_sales VALUE month_1_sales, month_2_sales, month_3_sales; ``` -
| empid | dept | quarter | month_1_sales | month_2_sales | month_3_sales | |------:|-------------|---------|--------------:|--------------:|--------------:| @@ -243,7 +239,6 @@ FROM monthly_sales; Note the single quotes to build a list of text strings to populate `month`, and the double quotes to pull the column values for use in `sales`. This produces the same result as the initial example: -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -297,7 +292,6 @@ FROM monthly_sales UNPIVOT ( ); ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -356,7 +350,6 @@ UNPIVOT ( ); ``` -
| empid | dept | quarter | month_1_sales | month_2_sales | month_3_sales | |------:|-------------|---------|--------------:|--------------:|--------------:| diff --git a/docs/archive/1.0/sql/statements/update.md b/docs/archive/1.0/sql/statements/update.md index 93fe8ced8ca..4846c2719a1 100644 --- a/docs/archive/1.0/sql/statements/update.md +++ b/docs/archive/1.0/sql/statements/update.md @@ -48,7 +48,6 @@ SELECT * FROM original; ``` -
| key | value | |-----|------------------| @@ -79,7 +78,6 @@ SELECT * FROM original; ``` -
| key | value | |-----|-------------| diff --git a/docs/configuration/pragmas.md b/docs/configuration/pragmas.md index d94522582e0..9b3043d11f1 100644 --- a/docs/configuration/pragmas.md +++ b/docs/configuration/pragmas.md @@ -100,7 +100,6 @@ CALL pragma_storage_info('table_name'); This call returns the following information for the given table: -
| Name | Type | Description | |----------------|-----------|-------------------------------------------------------| diff --git a/docs/data/csv/auto_detection.md b/docs/data/csv/auto_detection.md index ff1de330c1b..3a82ad38344 100644 --- a/docs/data/csv/auto_detection.md +++ b/docs/data/csv/auto_detection.md @@ -61,7 +61,6 @@ Dialect detection works by attempting to parse the samples using the set of cons The following dialects are considered for automatic dialect detection. -
diff --git a/docs/data/csv/overview.md b/docs/data/csv/overview.md index 04b02f7ff2e..ea5d822dead 100644 --- a/docs/data/csv/overview.md +++ b/docs/data/csv/overview.md @@ -136,7 +136,6 @@ The `read_csv` automatically attempts to figure out the correct configuration of SELECT * FROM read_csv('flights.csv'); ``` -
| FlightDate | UniqueCarrier | OriginCityName | DestCityName | |------------|---------------|----------------|-----------------| @@ -154,7 +153,6 @@ CREATE TABLE ontime AS DESCRIBE ontime; ``` -
| column_name | column_type | null | key | default | extra | |----------------|-------------|------|------|---------|-------| @@ -190,7 +188,6 @@ COPY ontime FROM 'flights.csv'; SELECT * FROM ontime; ``` -
| flightdate | uniquecarrier | origincityname | destcityname | |------------|---------------|----------------|-----------------| diff --git a/docs/data/csv/reading_faulty_csv_files.md b/docs/data/csv/reading_faulty_csv_files.md index c231ac22913..52f4fc29afc 100644 --- a/docs/data/csv/reading_faulty_csv_files.md +++ b/docs/data/csv/reading_faulty_csv_files.md @@ -112,7 +112,6 @@ FROM read_csv( Outputs: -
| name | age | |-------|-----| @@ -127,7 +126,6 @@ FROM read_csv('faulty.csv', columns = {'name': 'VARCHAR', 'age': 'INTEGER'}); Outputs: -
| name | |--------------| @@ -148,7 +146,6 @@ Note that any of the errors described in our Structural Error section will be st The CSV Reject Scans Table returns the following information: -
| Column name | Description | Type | |:--|:-----|:-| @@ -170,7 +167,6 @@ The CSV Reject Scans Table returns the following information: The CSV Reject Errors Table returns the following information: -
| Column name | Description | Type | |:--|:-----|:-| @@ -187,7 +183,6 @@ The CSV Reject Errors Table returns the following information: ## Parameters -
The parameters listed below are used in the `read_csv` function to configure the CSV Rejects Table. diff --git a/docs/data/json/creating_json.md b/docs/data/json/creating_json.md index 06a2e768695..9033b412695 100644 --- a/docs/data/json/creating_json.md +++ b/docs/data/json/creating_json.md @@ -7,7 +7,6 @@ title: Creating JSON The following functions are used to create JSON. -
| Function | Description | |:--|:----| diff --git a/docs/data/json/json_functions.md b/docs/data/json/json_functions.md index 6b18cd9a342..a15f262ef67 100644 --- a/docs/data/json/json_functions.md +++ b/docs/data/json/json_functions.md @@ -369,7 +369,6 @@ true There are three JSON aggregate functions. -
| Function | Description | |:---|:----| @@ -420,7 +419,6 @@ SELECT json_group_structure(j) FROM example2; In many cases, it is inefficient to extract values from JSON one-by-one. Instead, we can “extract” all values at once, transforming JSON to the nested types `LIST` and `STRUCT`. -
| Function | Description | |:---|:---| diff --git a/docs/data/multiple_files/combining_schemas.md b/docs/data/multiple_files/combining_schemas.md index 75c01584a14..a7464e63503 100644 --- a/docs/data/multiple_files/combining_schemas.md +++ b/docs/data/multiple_files/combining_schemas.md @@ -48,7 +48,6 @@ FlightDate|UniqueCarrier|OriginCityName|DestCityName Reading the two files at the same time will produce the following result set: -
| FlightDate | UniqueCarrier | OriginCityName | DestCityName | |------------|---------------|----------------|-----------------| @@ -83,7 +82,6 @@ Reading these when unifying column names **by position** results in an error – SELECT * FROM read_csv(['flights3.csv', 'flights4.csv'], union_by_name = true); ``` -
| FlightDate | OriginCityName | DestCityName | UniqueCarrier | |------------|----------------|-----------------|---------------| diff --git a/docs/data/multiple_files/overview.md b/docs/data/multiple_files/overview.md index 1ee93672d18..05528d567e2 100644 --- a/docs/data/multiple_files/overview.md +++ b/docs/data/multiple_files/overview.md @@ -94,7 +94,6 @@ FROM read_parquet(['file1.parquet', 'file2.parquet', 'file3.parquet']); Any file name input to the `read_parquet` function can either be an exact filename, or use a glob syntax to read multiple files that match a pattern. -
| Wildcard | Description | |------------|-----------------------------------------------------------| @@ -137,7 +136,6 @@ SELECT * FROM read_csv(['flights1.csv', 'flights2.csv'], union_by_name = true, filename = true); ``` -
| FlightDate | OriginCityName | DestCityName | UniqueCarrier | filename | |------------|----------------|-----------------|---------------|--------------| @@ -157,7 +155,6 @@ SELECT * FROM glob('*'); ``` -
| file | |---------------| diff --git a/docs/dev/release_calendar.md b/docs/dev/release_calendar.md index be4586776ec..89fccc6b068 100644 --- a/docs/dev/release_calendar.md +++ b/docs/dev/release_calendar.md @@ -18,7 +18,6 @@ Patch versions only ship bugfixes, while minor versions also introduce new featu The planned dates of upcoming DuckDB releases are shown below. **Please note that these dates are tentative** and DuckDB maintainers may decide to push back release dates to ensure the stability and quality of releases. -
@@ -36,7 +35,6 @@ _There are no upcoming releases announced at the moment. Please check back later ## Past Releases -
In the following, we list DuckDB's past releases along with their codename where applicable. Between versions 0.2.2 and 0.3.3, all releases (including patch versions) received a codename. diff --git a/docs/extensions/arrow.md b/docs/extensions/arrow.md index 39c99a2f2db..82deab411f1 100644 --- a/docs/extensions/arrow.md +++ b/docs/extensions/arrow.md @@ -18,7 +18,6 @@ LOAD arrow; ## Functions -
| Function | Type | Description | |--|----|-------| diff --git a/docs/extensions/autocomplete.md b/docs/extensions/autocomplete.md index 31649df39dd..d27c4c7f00c 100644 --- a/docs/extensions/autocomplete.md +++ b/docs/extensions/autocomplete.md @@ -13,7 +13,6 @@ For the behavior of the `autocomplete` extension, see the [documentation of the ## Functions -
| Function | Description | |:----------------------------------|:-----------------------------------------------------| @@ -28,7 +27,6 @@ FROM sql_auto_complete('SEL'); Returns: -
| suggestion | suggestion_start | |-------------|------------------| diff --git a/docs/extensions/core_extensions.md b/docs/extensions/core_extensions.md index 2d244744836..5b1095fbd64 100644 --- a/docs/extensions/core_extensions.md +++ b/docs/extensions/core_extensions.md @@ -37,7 +37,6 @@ redirect_from: Different DuckDB clients ship a different set of extensions. We summarize the main distributions in the table below. -
| Name | CLI (duckdb.org) | CLI (Homebrew) | Python | R | Java | Node.js | |------|------|------|---|---|---|---|---| diff --git a/docs/extensions/full_text_search.md b/docs/extensions/full_text_search.md index 0a715d12d09..6de576acf3d 100644 --- a/docs/extensions/full_text_search.md +++ b/docs/extensions/full_text_search.md @@ -56,7 +56,6 @@ drop_fts_index(input_table) Drops a FTS index for the specified table. -
| Name | Type | Description | |:--|:--|:-----------| diff --git a/docs/extensions/httpfs/s3api.md b/docs/extensions/httpfs/s3api.md index 1c15b04e423..0e0488c3398 100644 --- a/docs/extensions/httpfs/s3api.md +++ b/docs/extensions/httpfs/s3api.md @@ -11,7 +11,6 @@ The `httpfs` filesystem is tested with [AWS S3](https://aws.amazon.com/s3/), [Mi The following table shows which parts of the S3 API are required for each `httpfs` feature. -
| Feature | Required S3 API features | |:---|:---| @@ -199,7 +198,6 @@ FROM read_parquet('s3://bucket/*.parquet', filename = true); could for example result in: -
| column_a | column_b | filename | |:---|:---|:---| @@ -247,7 +245,6 @@ s3://my-bucket/partitioned/part_col_a=⟨val⟩/part_col_b=⟨val⟩/data_⟨thr Some additional configuration options exist for the S3 upload, though the default values should suffice for most use cases. -
| Name | Description | |:---|:---| diff --git a/docs/extensions/httpfs/s3api_legacy_authentication.md b/docs/extensions/httpfs/s3api_legacy_authentication.md index 0004ab0b739..b101d4b3d8f 100644 --- a/docs/extensions/httpfs/s3api_legacy_authentication.md +++ b/docs/extensions/httpfs/s3api_legacy_authentication.md @@ -73,7 +73,6 @@ Some additional configuration options exist for the S3 upload, though the defaul Additionally, most of the configuration options can be set via environment variables: -
| DuckDB setting | Environment variable | Note | |:-----------------------|:------------------------|:-----------------------------------------| diff --git a/docs/extensions/overview.md b/docs/extensions/overview.md index 232edfba308..64812c720a1 100644 --- a/docs/extensions/overview.md +++ b/docs/extensions/overview.md @@ -23,7 +23,6 @@ SELECT extension_name, installed, description FROM duckdb_extensions(); ``` -
| extension_name | installed | description | |-------------------|-----------|--------------------------------------------------------------| diff --git a/docs/extensions/tpch.md b/docs/extensions/tpch.md index a92e1f6b742..c42c21d749c 100644 --- a/docs/extensions/tpch.md +++ b/docs/extensions/tpch.md @@ -88,7 +88,6 @@ CALL dbgen(sf = 0); The data generator function `dbgen` has the following parameters: -
| Name | Type | Description | |--|--|------------| diff --git a/docs/extensions/versioning_of_extensions.md b/docs/extensions/versioning_of_extensions.md index 32152df9c27..61d78ca2a43 100644 --- a/docs/extensions/versioning_of_extensions.md +++ b/docs/extensions/versioning_of_extensions.md @@ -117,7 +117,6 @@ give the user information on which extensions were updated to/from which version UPDATE EXTENSIONS; ``` -
| extension_name | repository | update_result | previous_version | current_version | |:---------------|:-------------|:----------------------|:-----------------|:----------------| @@ -135,7 +134,6 @@ The update statement can also be provided with a list of specific extensions to UPDATE EXTENSIONS (httpfs, azure); ``` -
| extension_name | repository | update_result | previous_version | current_version | |:---------------|:-------------|:----------------------|:-----------------|:----------------| diff --git a/docs/extensions/working_with_extensions.md b/docs/extensions/working_with_extensions.md index f8fe5c45c02..36bf966d580 100644 --- a/docs/extensions/working_with_extensions.md +++ b/docs/extensions/working_with_extensions.md @@ -10,7 +10,6 @@ For platforms where packages for certain extensions are not available, users can All official extensions are distributed for the following platforms. -
| Platform name | Operating system | Architecture | CPU types | Used by | |--------------------|------------------|-----------------|--------------------------------|----------------------------| diff --git a/docs/guides/database_integration/mysql.md b/docs/guides/database_integration/mysql.md index 23c025d4dbb..4162219ad89 100644 --- a/docs/guides/database_integration/mysql.md +++ b/docs/guides/database_integration/mysql.md @@ -32,7 +32,6 @@ USE mysql_db; The string used by `ATTACH` is a PostgreSQL-style connection string (_not_ a MySQL connection string!). It is a list of connection arguments provided in `{key}={value}` format. Below is a list of valid arguments. Any options not provided are replaced by their default values. -
| Setting | Default | |------------|--------------| diff --git a/docs/guides/database_integration/postgres.md b/docs/guides/database_integration/postgres.md index 1515d52d048..77d029d29c0 100644 --- a/docs/guides/database_integration/postgres.md +++ b/docs/guides/database_integration/postgres.md @@ -32,7 +32,6 @@ SELECT * FROM postgres_scan('host=localhost port=5432 dbname=mydb', 'public', 'm The first parameter to the `postgres_scan` function is the [PostgreSQL connection string](https://www.postgresql.org/docs/current/libpq-connect.html#LIBPQ-CONNSTRING), a list of connection arguments provided in `{key}={value}` format. Below is a list of valid arguments. -
| Name | Description | Default | |------------|--------------------------------------|----------------| diff --git a/docs/guides/meta/describe.md b/docs/guides/meta/describe.md index 99969b7eabe..90c3e36a206 100644 --- a/docs/guides/meta/describe.md +++ b/docs/guides/meta/describe.md @@ -13,7 +13,6 @@ DESCRIBE tbl; SHOW tbl; -- equivalent to DESCRIBE tbl; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|------|---------|-------| @@ -28,7 +27,6 @@ In order to view the schema of the result of a query, prepend `DESCRIBE` to a qu DESCRIBE SELECT * FROM tbl; ``` -
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|------|---------|-------| diff --git a/docs/guides/meta/list_tables.md b/docs/guides/meta/list_tables.md index c7f1a465f10..4448afe17fa 100644 --- a/docs/guides/meta/list_tables.md +++ b/docs/guides/meta/list_tables.md @@ -10,7 +10,6 @@ CREATE TABLE tbl (i INTEGER); SHOW TABLES; ``` -
| name | |------| @@ -25,7 +24,6 @@ CREATE TABLE s1.tbl (v VARCHAR); SHOW ALL TABLES; ``` -
| database | schema | table_name | column_names | column_types | temporary | |----------|--------|------------|--------------|--------------|-----------| diff --git a/docs/guides/network_cloud_storage/s3_express_one.md b/docs/guides/network_cloud_storage/s3_express_one.md index 989df84e0ed..23e7a3f03c4 100644 --- a/docs/guides/network_cloud_storage/s3_express_one.md +++ b/docs/guides/network_cloud_storage/s3_express_one.md @@ -70,7 +70,6 @@ FROM 's3://express-bucket-name--use1-az5--x-s3/my-file.parquet'; We ran two experiments on a `c7gd.12xlarge` instance using the [LDBC SF300 Comments `creationDate` Parquet file](https://blobs.duckdb.org/data/ldbc-sf300-comments-creationDate.parquet) file (also used in the [microbenchmarks of the performance guide]({% link docs/guides/performance/benchmarks.md %}#data-sets)). -
| Experiment | File size | Runtime | |:-----|--:|--:| diff --git a/docs/guides/performance/file_formats.md b/docs/guides/performance/file_formats.md index 4e5d8cca3dc..398083de26d 100644 --- a/docs/guides/performance/file_formats.md +++ b/docs/guides/performance/file_formats.md @@ -40,7 +40,6 @@ DuckDB works best on Parquet files with row groups of 100K-1M rows each. The rea We run a simple aggregation query over Parquet files using different row group sizes, selected between 960 and 1,966,080. The results are as follows. -
| Row group size | Execution time | |---------------:|---------------:| @@ -77,7 +76,6 @@ For tips on reading and writing Parquet files, see the [Parquet Tips page]({% li CSV files are often distributed in compressed format such as GZIP archives (`.csv.gz`). DuckDB can decompress these files on the fly. In fact, this is typically faster than decompressing the files first and loading them due to reduced IO. -
| Schema | Load time | |---|--:| diff --git a/docs/guides/performance/indexing.md b/docs/guides/performance/indexing.md index 254d9695fbc..d11404ecbde 100644 --- a/docs/guides/performance/indexing.md +++ b/docs/guides/performance/indexing.md @@ -17,7 +17,6 @@ The more ordered the data within a column, the more useful the zonemap indexes w For an example, let’s repeat the [microbenchmark for timestamps]({% link docs/guides/performance/schema.md %}#microbenchmark-using-timestamps) with a timestamp column that sorted using an ascending order vs. an unordered one. -
| Column type | Ordered | Storage size | Query time | |---|---|--:|--:| diff --git a/docs/guides/performance/schema.md b/docs/guides/performance/schema.md index 50c8683d1f5..9cec23c2367 100644 --- a/docs/guides/performance/schema.md +++ b/docs/guides/performance/schema.md @@ -31,7 +31,6 @@ SELECT avg(CAST(creationDate[9:10] AS INTEGER)) FROM Comment; The results of the microbenchmark are as follows: -
| Column type | Storage size | Query time | | ----------- | -----------: | ---------: | @@ -55,7 +54,6 @@ In the second experiment, we define all columns with the `VARCHAR` type. While the results of the queries are the same for all both experiments, their runtime vary significantly. The results below show that joining on `BIGINT` columns is approx. 1.8× faster than performing the same join on `VARCHAR`-typed columns encoding the same value. -
| Join column payload type | Join column schema type | Example value | Query time | | ------------------------ | ----------------------- | ---------------------------------------- | ---------: | @@ -74,7 +72,6 @@ DuckDB allows defining [constraints]({% link docs/sql/constraints.md %}) such as We illustrate the effect of using primary keys with the [LDBC Comment table at scale factor 300](https://blobs.duckdb.org/data/ldbc-sf300-comments.tar.zst). This table has approx. 554 million entries. We first create the schema without a primary key, then load the data. In the second experiment, we create the schema with a primary key, then load the data. In both cases, we take the data from `.csv.gz` files, and measure the time required to perform the loading. -
| Operation | Execution time | | ------------------------ | -------------: | diff --git a/docs/guides/sql_features/asof_join.md b/docs/guides/sql_features/asof_join.md index acd4ad23207..17c95fdc6f6 100644 --- a/docs/guides/sql_features/asof_join.md +++ b/docs/guides/sql_features/asof_join.md @@ -24,7 +24,6 @@ which can be cumbersome and slow to implement in standard SQL. Let's start with a concrete example. Suppose we have a table of stock [`prices`](/data/prices.csv) with timestamps: -
| ticker | when | price | | :----- | :--- | ----: | @@ -40,7 +39,6 @@ Suppose we have a table of stock [`prices`](/data/prices.csv) with timestamps: We have another table containing portfolio [`holdings`](/data/holdings.csv) at various points in time: -
| ticker | when | shares | | :----- | :--- | -----: | @@ -76,7 +74,6 @@ ASOF JOIN prices p This attaches the value of the holding at that time to each row: -
| ticker | when | value | | :----- | :--- | ----: | @@ -107,7 +104,6 @@ ORDER BY ALL; As you might expect, this will produce `NULL` prices and values instead of dropping left side rows when there is no ticker or the time is before the prices begin. -
| ticker | when | value | | :----- | :--- | ----: | diff --git a/docs/internals/pivot.md b/docs/internals/pivot.md index 5eda3773588..4821070b75c 100644 --- a/docs/internals/pivot.md +++ b/docs/internals/pivot.md @@ -53,7 +53,6 @@ GROUP BY ALL; This produces the result: -
| country | name | list("year") | list(population_sum) | |---------|---------------|--------------------|----------------------| @@ -63,7 +62,6 @@ This produces the result: The `PhysicalPivot` operator converts those lists into column names and values to return this result: -
| country | name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -103,7 +101,6 @@ FROM monthly_sales; Note the single quotes to build a list of text strings to populate `month`, and the double quotes to pull the column values for use in `sales`. This produces the same result as the initial example: -
| empid | dept | month | sales | |------:|-------------|-------|------:| diff --git a/docs/internals/storage.md b/docs/internals/storage.md index c24c0da4d0f..ae92bb2d3b2 100644 --- a/docs/internals/storage.md +++ b/docs/internals/storage.md @@ -67,7 +67,6 @@ with open('test/sql/storage_version/storage_version.db', 'rb') as fh: For changes in each given release, check out the [change log](https://github.com/duckdb/duckdb/releases) on GitHub. To see the commits that changed each storage version, see the [commit log](https://github.com/duckdb/duckdb/commits/main/src/storage/storage_info.cpp). -
| Storage version | DuckDB version(s) | |----------------:|---------------------------------| diff --git a/docs/sql/data_types/bitstring.md b/docs/sql/data_types/bitstring.md index 88ccab8154e..280225d74c7 100644 --- a/docs/sql/data_types/bitstring.md +++ b/docs/sql/data_types/bitstring.md @@ -4,7 +4,6 @@ title: Bitstring Type blurb: The bitstring type are strings of 1s and 0s. --- -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/sql/data_types/blob.md b/docs/sql/data_types/blob.md index fb3f6357fc4..8a2b40bf9cf 100644 --- a/docs/sql/data_types/blob.md +++ b/docs/sql/data_types/blob.md @@ -4,7 +4,6 @@ title: Blob Type blurb: The blob (Binary Large OBject) type represents an arbitrary binary object stored in the database system. --- -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/sql/data_types/boolean.md b/docs/sql/data_types/boolean.md index 7f002bdb9b1..186c6e0af46 100644 --- a/docs/sql/data_types/boolean.md +++ b/docs/sql/data_types/boolean.md @@ -4,7 +4,6 @@ title: Boolean Type blurb: The BOOLEAN type represents a statement of truth (“true” or “false”). --- -
| Name | Aliases | Description | |:---|:---|:---| @@ -45,7 +44,6 @@ The `AND`/`OR` conjunctions can be used to combine Boolean values. Below is the truth table for the `AND` conjunction (i.e., `x AND y`). -
| X | X AND true | X AND false | X AND NULL | |-------|-------|-------|-------| @@ -55,7 +53,6 @@ Below is the truth table for the `AND` conjunction (i.e., `x AND y`). Below is the truth table for the `OR` conjunction (i.e., `x OR y`). -
| X | X OR true | X OR false | X OR NULL | |-------|------|-------|------| diff --git a/docs/sql/data_types/date.md b/docs/sql/data_types/date.md index 5f9e8b112f1..d7df80479f0 100644 --- a/docs/sql/data_types/date.md +++ b/docs/sql/data_types/date.md @@ -4,7 +4,6 @@ title: Date Types blurb: A date specifies a combination of year, month and day. --- -
| Name | Aliases | Description | |:-------|:--------|:--------------------------------| @@ -20,7 +19,6 @@ SELECT DATE '1992-09-20'; There are also three special date values that can be used on input: -
| Input string | Description | |:-------------|:----------------------------------| diff --git a/docs/sql/data_types/enum.md b/docs/sql/data_types/enum.md index 322ad8d0ca9..1860591a6c5 100644 --- a/docs/sql/data_types/enum.md +++ b/docs/sql/data_types/enum.md @@ -4,7 +4,6 @@ title: Enum Data Type blurb: The Enum type represents a dictionary data structure with all possible unique values of a column. --- -
| Name | Description | |:--|:-----| @@ -75,7 +74,6 @@ Show the available values in the `birds` enum using the `enum_range` function: SELECT enum_range(NULL::birds) AS my_enum_range; ``` -
| my_enum_range | |-----------------| diff --git a/docs/sql/data_types/interval.md b/docs/sql/data_types/interval.md index 5d38693a44c..35f6c8233c7 100644 --- a/docs/sql/data_types/interval.md +++ b/docs/sql/data_types/interval.md @@ -6,7 +6,6 @@ blurb: Intervals represent periods of time measured in months, days, microsecond `INTERVAL`s represent periods of time that can be added to or subtracted from `DATE`, `TIMESTAMP`, `TIMESTAMPTZ`, or `TIME` values. -
| Name | Description | |:---|:---| diff --git a/docs/sql/data_types/numeric.md b/docs/sql/data_types/numeric.md index 748a672169b..80a6fff80c5 100644 --- a/docs/sql/data_types/numeric.md +++ b/docs/sql/data_types/numeric.md @@ -34,7 +34,6 @@ Division of fixed-point decimals does not typically produce numbers with finite Internally, decimals are represented as integers depending on their specified `WIDTH`. -
| Width | Internal | Size (bytes) | |:---|:---|---:| @@ -49,7 +48,6 @@ Performance can be impacted by using too large decimals when not required. In pa The data types `FLOAT` and `DOUBLE` precision are variable-precision numeric types. In practice, these types are usually implementations of IEEE Standard 754 for Binary Floating-Point Arithmetic (single and double precision, respectively), to the extent that the underlying processor, operating system, and compiler support it. -
| Name | Aliases | Description | |:--|:--|:--------| diff --git a/docs/sql/data_types/struct.md b/docs/sql/data_types/struct.md index 65f4c9e5cd1..893a1341770 100644 --- a/docs/sql/data_types/struct.md +++ b/docs/sql/data_types/struct.md @@ -111,7 +111,6 @@ SELECT a.* FROM (SELECT {'x': 1, 'y': 2, 'z': 3} AS a); ``` -
| x | y | z | |:---|:---|:---| diff --git a/docs/sql/data_types/text.md b/docs/sql/data_types/text.md index 0145b0ea649..8299b9a01f0 100644 --- a/docs/sql/data_types/text.md +++ b/docs/sql/data_types/text.md @@ -7,7 +7,6 @@ blurb: In DuckDB, strings can be stored in the VARCHAR field. In DuckDB, strings can be stored in the `VARCHAR` field. The field allows storage of Unicode characters. Internally, the data is encoded as UTF-8. -
| Name | Aliases | Description | |:---|:---|:---| diff --git a/docs/sql/data_types/time.md b/docs/sql/data_types/time.md index d3e3df89bc0..9b4ae6c1a4a 100644 --- a/docs/sql/data_types/time.md +++ b/docs/sql/data_types/time.md @@ -6,7 +6,6 @@ blurb: A time instance represents the time of a day (hour, minute, second, micro The `TIME` and `TIMETZ` types specify the hour, minute, second, microsecond of a day. -
| Name | Aliases | Description | | :------- | :----------------------- | :------------------------------ | diff --git a/docs/sql/data_types/timestamp.md b/docs/sql/data_types/timestamp.md index 56924dbba71..9141d9e9fd7 100644 --- a/docs/sql/data_types/timestamp.md +++ b/docs/sql/data_types/timestamp.md @@ -81,7 +81,6 @@ SELECT TIMESTAMP WITH TIME ZONE '1992-09-20 11:30:00.123456789'; There are also three special date values that can be used on input: -
| Input string | Valid types | Description | |:-------------|:---------------------------|:-----------------------------------------------| @@ -96,7 +95,6 @@ but `epoch` is simply a notational shorthand that will be converted to the time SELECT '-infinity'::TIMESTAMP, 'epoch'::TIMESTAMP, 'infinity'::TIMESTAMP; ``` -
| Negative | Epoch | Positive | |:----------|:--------------------|:---------| diff --git a/docs/sql/data_types/timezones.md b/docs/sql/data_types/timezones.md index 4bbdbb10d25..d12e5d3c706 100644 --- a/docs/sql/data_types/timezones.md +++ b/docs/sql/data_types/timezones.md @@ -12,7 +12,6 @@ FROM pg_timezone_names() ORDER BY name; ``` -
| name | abbrev | |----------------------------------|----------------------------------| diff --git a/docs/sql/data_types/union.md b/docs/sql/data_types/union.md index d8a743fec82..ad29083b5fb 100644 --- a/docs/sql/data_types/union.md +++ b/docs/sql/data_types/union.md @@ -95,7 +95,6 @@ So how do we disambiguate if we want to create a `UNION` with multiple members o `UNION` types can be cast between each other if the source type is a “subset” of the target type. In other words, all the tags in the source `UNION` must be present in the target `UNION`, and all the types of the matching tags must be implicitly castable between source and target. In essence, this means that `UNION` types are covariant with respect to their members. -
| Ok | Source | Target | Comments | |----|------------------------|------------------------|----------------------------------------| diff --git a/docs/sql/dialect/postgresql_compatibility.md b/docs/sql/dialect/postgresql_compatibility.md index 6b5c3895fb7..c3022d305af 100644 --- a/docs/sql/dialect/postgresql_compatibility.md +++ b/docs/sql/dialect/postgresql_compatibility.md @@ -44,7 +44,6 @@ SELECT 1 / 2 AS x; PostgreSQL returns: -
| x | | ---: | @@ -52,7 +51,6 @@ PostgreSQL returns: DuckDB returns: -
| x | | ---: | @@ -64,7 +62,6 @@ To perform integer division in DuckDB, use the `//` operator: SELECT 1 // 2 AS x; ``` -
| x | | ---: | @@ -89,7 +86,6 @@ ERROR: UNION types boolean and integer cannot be matched DuckDB performs an enforced cast, therefore, it completes the query and returns the following: -
| x | | ---: | diff --git a/docs/sql/expressions/comparison_operators.md b/docs/sql/expressions/comparison_operators.md index a6f7ab632aa..866a4ea20c6 100644 --- a/docs/sql/expressions/comparison_operators.md +++ b/docs/sql/expressions/comparison_operators.md @@ -11,7 +11,6 @@ railroad: expressions/comparison.js The table below shows the standard comparison operators. Whenever either of the input arguments is `NULL`, the output of the comparison is `NULL`. -
| Operator | Description | Example | Result | |:---|:---|:---|:---| @@ -25,7 +24,6 @@ Whenever either of the input arguments is `NULL`, the output of the comparison i The table below shows the standard distinction operators. These operators treat `NULL` values as equal. -
| Operator | Description | Example | Result | |:---|:---|:---|:-| @@ -40,7 +38,6 @@ Besides the standard comparison operators there are also the `BETWEEN` and `IS ( Note that `BETWEEN` and `NOT BETWEEN` are only equivalent to the examples below in the cases where both `a`, `x` and `y` are of the same type, as `BETWEEN` will cast all of its inputs to the same type. -
| Predicate | Description | |:---|:---| diff --git a/docs/sql/expressions/logical_operators.md b/docs/sql/expressions/logical_operators.md index 891b32d743d..593398a6747 100644 --- a/docs/sql/expressions/logical_operators.md +++ b/docs/sql/expressions/logical_operators.md @@ -10,7 +10,6 @@ The following logical operators are available: `AND`, `OR` and `NOT`. SQL uses a ### Binary Operators: `AND` and `OR` -
| `a` | `b` | `a AND b` | `a OR b` | |:---|:---|:---|:---| @@ -23,7 +22,6 @@ The following logical operators are available: `AND`, `OR` and `NOT`. SQL uses a ### Unary Operator: `NOT` -
| `a` | `NOT a` | |:---|:---| diff --git a/docs/sql/expressions/star.md b/docs/sql/expressions/star.md index 3b7a5e49feb..a80f94df679 100644 --- a/docs/sql/expressions/star.md +++ b/docs/sql/expressions/star.md @@ -110,7 +110,6 @@ INSERT INTO numbers VALUES (1, 10), (2, 20), (3, NULL); SELECT min(COLUMNS(*)), count(COLUMNS(*)) FROM numbers; ``` -
| id | number | id | number | |---:|-------:|---:|-------:| @@ -125,7 +124,6 @@ SELECT FROM numbers; ``` -
| id | min(number := (number + id)) | id | |---:|-----------------------------:|---:| @@ -137,7 +135,6 @@ FROM numbers; SELECT COLUMNS(*) + COLUMNS(*) FROM numbers; ``` -
| id | number | |---:|-------:| @@ -159,7 +156,6 @@ FROM ( WHERE COLUMNS(*) > 1; -- equivalent to: x > 1 AND y > 1 AND z > 1 ``` -
| x | y | z | |--:|--:|--:| @@ -173,7 +169,6 @@ WHERE COLUMNS(*) > 1; -- equivalent to: x > 1 AND y > 1 AND z > 1 SELECT COLUMNS('(id|numbers?)') FROM numbers; ``` -
| id | number | |---:|-------:| @@ -192,7 +187,6 @@ For example, to select the first three letters of colum names, run: SELECT COLUMNS('(\w{3}).*') AS '\1' FROM numbers; ``` -
| id | num | |---:|-----:| @@ -215,7 +209,6 @@ SELECT COLUMNS('(\w*):(\w*)') AS '\1\2' FROM tbl; SELECT COLUMNS(c -> c LIKE '%num%') FROM numbers; ``` -
| number | |-------:| @@ -277,7 +270,6 @@ For example: SELECT st.* FROM (SELECT {'x': 1, 'y': 2, 'z': 3} AS st); ``` -
| x | y | z | |--:|--:|--:| diff --git a/docs/sql/expressions/subqueries.md b/docs/sql/expressions/subqueries.md index 3231935b932..2db88e073f5 100644 --- a/docs/sql/expressions/subqueries.md +++ b/docs/sql/expressions/subqueries.md @@ -16,7 +16,6 @@ Consider the following table: ### Grades -
| grade | course | |---:|:---| diff --git a/docs/sql/functions/bitstring.md b/docs/sql/functions/bitstring.md index ba63b8468d2..139a5a9ae02 100644 --- a/docs/sql/functions/bitstring.md +++ b/docs/sql/functions/bitstring.md @@ -14,7 +14,6 @@ Bitstrings must be of equal length when performing the bitwise operands AND, OR The table below shows the available mathematical operators for `BIT` type. -
diff --git a/docs/sql/functions/char.md b/docs/sql/functions/char.md index 7caf23e5148..5054aa0a551 100644 --- a/docs/sql/functions/char.md +++ b/docs/sql/functions/char.md @@ -883,7 +883,6 @@ SELECT format('I''d rather be {1} than {0}.', 'right', 'happy'); -- I'd rather b #### Format Specifiers -
| Specifier | Description | Example | |:-|:------|:---| @@ -1004,7 +1003,6 @@ I'd rather be happy than right. #### Format Specifiers -
| Specifier | Description | Example | |:-|:------|:---| diff --git a/docs/sql/functions/dateformat.md b/docs/sql/functions/dateformat.md index c080864fd82..d0eaf35cb03 100644 --- a/docs/sql/functions/dateformat.md +++ b/docs/sql/functions/dateformat.md @@ -85,7 +85,6 @@ FROM read_csv('test.csv', dateformat = '%m/%d/%Y'); Below is a full list of all available format specifiers. -
| Specifier | Description | Example | |:-|:------|:---| diff --git a/docs/sql/functions/numeric.md b/docs/sql/functions/numeric.md index b25054dbab3..12477313d00 100644 --- a/docs/sql/functions/numeric.md +++ b/docs/sql/functions/numeric.md @@ -11,7 +11,6 @@ redirect_from: The table below shows the available mathematical operators for [numeric types]({% link docs/sql/data_types/numeric.md %}). -
diff --git a/docs/sql/functions/pattern_matching.md b/docs/sql/functions/pattern_matching.md index 0f1e2c5d6a1..a76594b2ecd 100644 --- a/docs/sql/functions/pattern_matching.md +++ b/docs/sql/functions/pattern_matching.md @@ -144,7 +144,6 @@ SELECT NOT 'best.txt' GLOB '*.txt'; -- false Three tildes (`~~~`) may also be used in place of the `GLOB` keyword. -
| GLOB-style | Symbolic-style | |:---|:---| diff --git a/docs/sql/functions/regular_expressions.md b/docs/sql/functions/regular_expressions.md index 64286d24cf1..1c61f4afa69 100644 --- a/docs/sql/functions/regular_expressions.md +++ b/docs/sql/functions/regular_expressions.md @@ -113,7 +113,6 @@ SELECT regexp_matches('abc', 'A', 'i'); -- true The regex functions support the following `options`. -
| Option | Description | |:---|:---| @@ -138,7 +137,6 @@ SELECT regexp_matches(E'hello\nworld', 'hello.world', 's'); -- true The `regexp_matches` operator will be optimized to the `LIKE` operator when possible. To achieve best performance, the `'c'` option (case-sensitive matching) should be passed if applicable. Note that by default the [`RE2` library](#regular-expression-syntax) doesn't match the `.` character to newline. -
| Original | Optimized equivalent | |:---|:---| diff --git a/docs/sql/functions/time.md b/docs/sql/functions/time.md index 862c8bdef79..001cb4b4066 100644 --- a/docs/sql/functions/time.md +++ b/docs/sql/functions/time.md @@ -11,7 +11,6 @@ This section describes functions and operators for examining and manipulating [` The table below shows the available mathematical operators for `TIME` types. -
| Operator | Description | Example | Result | |:-|:---|:----|:--| diff --git a/docs/sql/functions/utility.md b/docs/sql/functions/utility.md index 1f5fb68e300..aa37a819ca0 100644 --- a/docs/sql/functions/utility.md +++ b/docs/sql/functions/utility.md @@ -357,7 +357,6 @@ The functions below are difficult to categorize into specific function types and A table function is used in place of a table in a `FROM` clause. -
| Name | Description | |:--|:-------| diff --git a/docs/sql/functions/window_functions.md b/docs/sql/functions/window_functions.md index 4f55d375c7a..6b3a60ab943 100644 --- a/docs/sql/functions/window_functions.md +++ b/docs/sql/functions/window_functions.md @@ -222,7 +222,6 @@ CREATE TABLE "Generation History" AS After partitioning by plant and ordering by date, it will have this layout: -
| Plant | Date | MWh | |:---|:---|---:| @@ -268,7 +267,6 @@ ORDER BY 1, 2; The result will be the following: -
| Plant | Date | Row | |:---|:---|---:| @@ -340,7 +338,6 @@ and uses a `RANGE` frame of three days on either side of each day for the `avg` (to handle any missing days). This is the result: -
| Plant | Date | MWh 7-day Moving Average | |:---|:---|---:| diff --git a/docs/sql/meta/duckdb_table_functions.md b/docs/sql/meta/duckdb_table_functions.md index 0959b9004a4..fc6c87827d5 100644 --- a/docs/sql/meta/duckdb_table_functions.md +++ b/docs/sql/meta/duckdb_table_functions.md @@ -170,7 +170,6 @@ Note that `duckdb_indexes` only provides metadata about secondary indexes, i.e., The `duckdb_keywords()` function provides metadata about DuckDB's keywords and reserved words. -
| Column | Description | Type | |:-|:---|:-| @@ -192,7 +191,6 @@ The `duckdb_memory()` function provides metadata about DuckDB's buffer manager. The `duckdb_optimizers()` function provides metadata about the optimization rules (e.g., `expression_rewriter`, `filter_pushdown`) available in the DuckDB instance. These can be selectively turned off using [`PRAGMA disabled_optimizers`]({% link docs/configuration/pragmas.md %}#selectively-disabling-optimizers). -
| Column | Description | Type | |:-|:---|:-| @@ -258,7 +256,6 @@ Attributes like `temporary`, `start_value` etc. correspond to the various option The `duckdb_settings()` function provides metadata about the settings available in the DuckDB instance. -
| Column | Description | Type | |:-|:---|:-| @@ -296,7 +293,6 @@ The [`information_schema.tables`]({% link docs/sql/meta/information_schema.md %} The `duckdb_temporary_files()` function provides metadata about the temporary files DuckDB has written to disk, to offload data from memory. This function mostly exists for debugging and testing purposes. -
| Column | Description | Type | |:-|:---|:-| diff --git a/docs/sql/query_syntax/orderby.md b/docs/sql/query_syntax/orderby.md index d12a79565d8..227d4148778 100644 --- a/docs/sql/query_syntax/orderby.md +++ b/docs/sql/query_syntax/orderby.md @@ -102,7 +102,6 @@ FROM addresses ORDER BY ALL; ``` -
| address | city | zip | |------------------------|-----------|------------| @@ -119,7 +118,6 @@ FROM addresses ORDER BY ALL DESC; ``` -
| address | city | zip | |------------------------|-----------|------------| diff --git a/docs/sql/query_syntax/qualify.md b/docs/sql/query_syntax/qualify.md index f029b117112..2ca6373dbc3 100644 --- a/docs/sql/query_syntax/qualify.md +++ b/docs/sql/query_syntax/qualify.md @@ -88,7 +88,6 @@ WHERE function_rank < 3; ``` -
| schema_name | function_name | function_rank | |:---|:---|:---| diff --git a/docs/sql/statements/attach.md b/docs/sql/statements/attach.md index 5c64fd84ae8..aab9194f243 100644 --- a/docs/sql/statements/attach.md +++ b/docs/sql/statements/attach.md @@ -121,7 +121,6 @@ USE memory_db; ## Options -
| Name | Description | Type | Default value | |-----------------------------|-----------------------------------------------------------------------------------------------------------------------------|-----------|---------------| diff --git a/docs/sql/statements/create_index.md b/docs/sql/statements/create_index.md index 63ba58a2b38..d40e6de9306 100644 --- a/docs/sql/statements/create_index.md +++ b/docs/sql/statements/create_index.md @@ -38,7 +38,6 @@ CREATE INDEX i_index ON integers ((j + k)); ### Parameters -
| Name | Description | |:-|:-----| @@ -68,7 +67,6 @@ DROP INDEX title_idx; ### Parameters -
| Name | Description | |:---|:---| diff --git a/docs/sql/statements/insert.md b/docs/sql/statements/insert.md index b04c59dcd27..96461efbdf9 100644 --- a/docs/sql/statements/insert.md +++ b/docs/sql/statements/insert.md @@ -388,7 +388,6 @@ INSERT INTO t1 RETURNING *; ``` -
| i | |---:| @@ -403,7 +402,6 @@ INSERT INTO t2 RETURNING *, i * j AS i_times_j; ``` -
| i | j | i_times_j | |--:|--:|----------:| @@ -421,7 +419,6 @@ INSERT INTO t3 RETURNING *; ``` -
| i | j | |--:|---:| diff --git a/docs/sql/statements/pivot.md b/docs/sql/statements/pivot.md index 9541287069e..fe49a0d9a77 100644 --- a/docs/sql/statements/pivot.md +++ b/docs/sql/statements/pivot.md @@ -55,7 +55,6 @@ SELECT * FROM cities; ``` -
| country | name | year | population | |---------|---------------|-----:|-----------:| @@ -85,7 +84,6 @@ ON year USING sum(population); ``` -
| country | name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -124,7 +122,6 @@ USING sum(population) GROUP BY country; ``` -
| country | 2000 | 2010 | 2020 | |---------|-----:|-----:|-----:| @@ -143,7 +140,6 @@ USING sum(population) GROUP BY country; ``` -
| country | 2000 | 2010 | |---------|-----:|-----:| @@ -168,7 +164,6 @@ ON country, name USING sum(population); ``` -
| year | NL_Amsterdam | NL_New York City | NL_Seattle | US_Amsterdam | US_New York City | US_Seattle | |-----:|-------------:|------------------|------------|--------------|-----------------:|-----------:| @@ -189,7 +184,6 @@ ON country || '_' || name USING sum(population); ``` -
| year | NL_Amsterdam | US_New York City | US_Seattle | |-----:|-------------:|-----------------:|-----------:| @@ -212,7 +206,6 @@ USING sum(population) AS total, max(population) AS max GROUP BY country; ``` -
| country | 2000_total | 2000_max | 2010_total | 2010_max | 2020_total | 2020_max | |---------|-----------:|---------:|-----------:|---------:|-----------:|---------:| @@ -231,7 +224,6 @@ USING sum(population) GROUP BY country, name; ``` -
| country | name | 2000 | 2010 | 2020 | |---------|---------------|-----:|-----:|-----:| @@ -282,7 +274,6 @@ JOIN (PIVOT cities ON name USING sum(population) GROUP BY country) name_pivot USING (country); ``` -
| country | 2000 | 2010 | 2020 | Amsterdam | New York City | Seattle | |---------|-----:|-----:|-----:|----------:|--------------:|--------:| @@ -332,7 +323,6 @@ PIVOT ( ); ``` -
| country | 2000 | 2010 | 2020 | |---------|-----:|-----:|-----:| diff --git a/docs/sql/statements/unpivot.md b/docs/sql/statements/unpivot.md index 812fc4fb9e5..7c3a51a17a5 100644 --- a/docs/sql/statements/unpivot.md +++ b/docs/sql/statements/unpivot.md @@ -47,7 +47,6 @@ INSERT INTO monthly_sales VALUES FROM monthly_sales; ``` -
| empid | dept | Jan | Feb | Mar | Apr | May | Jun | |------:|-------------|----:|----:|----:|----:|----:|----:| @@ -73,7 +72,6 @@ INTO VALUE sales; ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -112,7 +110,6 @@ INTO VALUE sales; ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -154,7 +151,6 @@ UNPIVOT monthly_sales VALUE month_1_sales, month_2_sales, month_3_sales; ``` -
| empid | dept | quarter | month_1_sales | month_2_sales | month_3_sales | |------:|-------------|---------|--------------:|--------------:|--------------:| @@ -246,7 +242,6 @@ FROM monthly_sales UNPIVOT ( ); ``` -
| empid | dept | month | sales | |------:|-------------|-------|------:| @@ -305,7 +300,6 @@ UNPIVOT ( ); ``` -
| empid | dept | quarter | month_1_sales | month_2_sales | month_3_sales | |------:|-------------|---------|--------------:|--------------:|--------------:| diff --git a/docs/sql/statements/update.md b/docs/sql/statements/update.md index 6abb3e59838..7849197657e 100644 --- a/docs/sql/statements/update.md +++ b/docs/sql/statements/update.md @@ -48,7 +48,6 @@ SELECT * FROM original; ``` -
| key | value | |-----|------------------| @@ -79,7 +78,6 @@ SELECT * FROM original; ``` -
| key | value | |-----|-------------| From 4e4443426816511dfdc32b41466e76dbd1c6b7ec Mon Sep 17 00:00:00 2001 From: Gabor Szarnyas Date: Sun, 8 Dec 2024 21:57:40 +0100 Subject: [PATCH 3/4] Remove narrow_table from multi-class divs --- docs/api/c/vector.md | 6 +++--- docs/api/python/types.md | 6 +++--- docs/archive/1.0/api/c/vector.md | 6 +++--- docs/archive/1.0/api/python/types.md | 6 +++--- docs/archive/1.0/data/csv/auto_detection.md | 8 ++++---- .../1.0/data/csv/reading_faulty_csv_files.md | 4 ++-- docs/archive/1.0/data/json/overview.md | 18 +++++++++--------- docs/archive/1.0/data/parquet/metadata.md | 8 ++++---- docs/archive/1.0/dev/sqllogictest/loops.md | 2 +- docs/archive/1.0/extensions/json.md | 16 ++++++++-------- docs/archive/1.0/extensions/mysql.md | 6 +++--- docs/archive/1.0/extensions/postgres.md | 4 ++-- docs/archive/1.0/extensions/spatial.md | 2 +- docs/archive/1.0/extensions/sqlite.md | 2 +- .../1.0/guides/meta/duckdb_environment.md | 4 ++-- docs/archive/1.0/sql/data_types/bitstring.md | 4 ++-- docs/archive/1.0/sql/data_types/list.md | 2 +- docs/archive/1.0/sql/data_types/numeric.md | 2 +- .../sql/dialect/postgresql_compatibility.md | 6 +++--- docs/archive/1.0/sql/expressions/subqueries.md | 2 +- .../1.0/sql/functions/pattern_matching.md | 8 ++++---- docs/archive/1.0/sql/query_syntax/filter.md | 10 +++++----- docs/data/csv/auto_detection.md | 8 ++++---- docs/data/csv/reading_faulty_csv_files.md | 4 ++-- docs/data/json/caveats.md | 2 +- docs/data/json/format_settings.md | 12 ++++++------ docs/data/json/json_functions.md | 2 +- docs/data/json/loading_json.md | 14 +++++++------- docs/data/parquet/metadata.md | 8 ++++---- docs/dev/sqllogictest/loops.md | 2 +- docs/extensions/aws.md | 6 +++--- docs/extensions/iceberg.md | 4 ++-- docs/extensions/mysql.md | 6 +++--- docs/extensions/postgres.md | 4 ++-- docs/extensions/sqlite.md | 2 +- docs/extensions/working_with_extensions.md | 2 +- docs/guides/meta/duckdb_environment.md | 4 ++-- docs/guides/sql_features/full_text_search.md | 2 +- docs/sql/data_types/bitstring.md | 4 ++-- docs/sql/data_types/list.md | 2 +- docs/sql/data_types/numeric.md | 2 +- docs/sql/dialect/postgresql_compatibility.md | 10 +++++----- docs/sql/expressions/subqueries.md | 2 +- docs/sql/functions/list.md | 6 +++--- docs/sql/functions/pattern_matching.md | 8 ++++---- docs/sql/query_syntax/filter.md | 10 +++++----- 46 files changed, 129 insertions(+), 129 deletions(-) diff --git a/docs/api/c/vector.md b/docs/api/c/vector.md index 9afd62c7630..fb67d109b8a 100644 --- a/docs/api/c/vector.md +++ b/docs/api/c/vector.md @@ -17,7 +17,7 @@ Vectors themselves do not have sizes. Instead, the parent data chunk has a size For primitive types, the underlying array can be obtained using the `duckdb_vector_get_data` method. The array can then be accessed using the correct native type. Below is a table that contains a mapping of the `duckdb_type` to the native type of the array. -
+
| duckdb_type | NativeType | |--------------------------|------------------| @@ -81,7 +81,7 @@ The length can either be accessed directly, or the `duckdb_string_is_inlined` ca Decimals are stored as integer values internally. The exact native type depends on the `width` of the decimal type, as shown in the following table: -
+
| Width | NativeType | |-------|----------------| @@ -98,7 +98,7 @@ Decimals are stored as integer values multiplied by `10^scale`. The scale of a d Enums are stored as unsigned integer values internally. The exact native type depends on the size of the enum dictionary, as shown in the following table: -
+
| Dictionary size | NativeType | |-----------------|------------| diff --git a/docs/api/python/types.md b/docs/api/python/types.md index ef0e58a78cc..1dacbeef3a5 100644 --- a/docs/api/python/types.md +++ b/docs/api/python/types.md @@ -14,7 +14,7 @@ This means that wherever a DuckDBPyType object is expected, it is also possible The table below shows the mapping of Python Built-in types to DuckDB type. -
+
| Built-in types | DuckDB type | |:---------------|:------------| @@ -29,7 +29,7 @@ The table below shows the mapping of Python Built-in types to DuckDB type. The table below shows the mapping of Numpy DType to DuckDB type. -
+
| Type | DuckDB type | |:------------|:------------| @@ -110,7 +110,7 @@ UNION(u1 BIGINT, u2 VARCHAR, u3 BOOLEAN, u4 BLOB) For the built-in types, you can use the constants defined in `duckdb.typing`: -
+
| DuckDB type | |:---------------| diff --git a/docs/archive/1.0/api/c/vector.md b/docs/archive/1.0/api/c/vector.md index 10ba711b548..795929478d9 100644 --- a/docs/archive/1.0/api/c/vector.md +++ b/docs/archive/1.0/api/c/vector.md @@ -17,7 +17,7 @@ Vectors themselves do not have sizes. Instead, the parent data chunk has a size For primitive types, the underlying array can be obtained using the `duckdb_vector_get_data` method. The array can then be accessed using the correct native type. Below is a table that contains a mapping of the `duckdb_type` to the native type of the array. -
+
| duckdb_type | NativeType | |--------------------------|------------------| @@ -81,7 +81,7 @@ The length can either be accessed directly, or the `duckdb_string_is_inlined` ca Decimals are stored as integer values internally. The exact native type depends on the `width` of the decimal type, as shown in the following table: -
+
| Width | NativeType | |-------|----------------| @@ -98,7 +98,7 @@ Decimals are stored as integer values multiplied by `10^scale`. The scale of a d Enums are stored as unsigned integer values internally. The exact native type depends on the size of the enum dictionary, as shown in the following table: -
+
| Dictionary Size | NativeType | |-----------------|------------| diff --git a/docs/archive/1.0/api/python/types.md b/docs/archive/1.0/api/python/types.md index 9bac824157b..7d75f602e11 100644 --- a/docs/archive/1.0/api/python/types.md +++ b/docs/archive/1.0/api/python/types.md @@ -14,7 +14,7 @@ This means that wherever a DuckDBPyType object is expected, it is also possible The table below shows the mapping of Python Built-in types to DuckDB type. -
+
| Built-in types | DuckDB type | |:---------------|:------------| @@ -29,7 +29,7 @@ The table below shows the mapping of Python Built-in types to DuckDB type. The table below shows the mapping of Numpy DType to DuckDB type. -
+
| Type | DuckDB type | |:------------|:------------| @@ -110,7 +110,7 @@ UNION(u1 BIGINT, u2 VARCHAR, u3 BOOLEAN, u4 BLOB) For the built-in types, you can use the constants defined in `duckdb.typing`: -
+
| DuckDB type | |:---------------| diff --git a/docs/archive/1.0/data/csv/auto_detection.md b/docs/archive/1.0/data/csv/auto_detection.md index 03d490489dc..955559fa519 100644 --- a/docs/archive/1.0/data/csv/auto_detection.md +++ b/docs/archive/1.0/data/csv/auto_detection.md @@ -95,7 +95,7 @@ After detecting the dialect, the system will attempt to figure out the types of The type detection works by attempting to convert the values in each column to the candidate types. If the conversion is unsuccessful, the candidate type is removed from the set of candidate types for that column. After all samples have been handled – the remaining candidate type with the highest priority is chosen. The default set of candidate types is given below, in order of priority: -
+
| Types | |-----------| @@ -113,7 +113,7 @@ The set of candidate types that should be considered by the CSV reader can be ex In addition to the default set of candidate types, other types that may be specified using the `auto_type_candidates` options are: -
+
| Types | |-----------| @@ -156,7 +156,7 @@ If the ambiguities cannot be resolved by looking at the data the system has a li The system considers the following formats for dates (`dateformat`). Higher entries are chosen over lower entries in case of ambiguities (i.e., ISO 8601 is preferred over `MM-DD-YYYY`). -
+
| dateformat | |------------| @@ -170,7 +170,7 @@ The system considers the following formats for dates (`dateformat`). Higher entr The system considers the following formats for timestamps (`timestampformat`). Higher entries are chosen over lower entries in case of ambiguities. -
+
| timestampformat | |----------------------| diff --git a/docs/archive/1.0/data/csv/reading_faulty_csv_files.md b/docs/archive/1.0/data/csv/reading_faulty_csv_files.md index f4339847ad6..524bd35ac79 100644 --- a/docs/archive/1.0/data/csv/reading_faulty_csv_files.md +++ b/docs/archive/1.0/data/csv/reading_faulty_csv_files.md @@ -211,7 +211,7 @@ FROM reject_scans; Outputs: -
+
| scan_id | file_id | file_path | delimiter | quote | escape | newline_delimiter | skip_rows | has_header | columns | date_format | timestamp_format | user_arguments | |---------|---------|-----------------------------------|-----------|-------|--------|-------------------|-----------|-----------:|--------------------------------------|-------------|------------------|--------------------| @@ -223,7 +223,7 @@ FROM reject_errors; Outputs: -
+
| scan_id | file_id | line | line_byte_position | byte_position | column_idx | column_name | error_type | csv_line | error_message | |---------|---------|------|--------------------|---------------|------------|-------------|------------|---------------------|------------------------------------------------------------------------------------| diff --git a/docs/archive/1.0/data/json/overview.md b/docs/archive/1.0/data/json/overview.md index ba2b74e1a30..172dded3c7e 100644 --- a/docs/archive/1.0/data/json/overview.md +++ b/docs/archive/1.0/data/json/overview.md @@ -111,7 +111,7 @@ SELECT * FROM read_json('records.json', format = 'newline_delimited'); ``` -
+
| key1 | key2 | |--------|--------| @@ -137,7 +137,7 @@ SELECT * FROM read_json('records-in-array.json', format = 'array'); ``` -
+
| key1 | key2 | |--------|--------| @@ -170,7 +170,7 @@ SELECT * FROM read_json('unstructured.json', format = 'unstructured'); ``` -
+
| key1 | key2 | |--------|--------| @@ -196,7 +196,7 @@ SELECT * FROM read_json('records.json', records = true); ``` -
+
| key1 | key2 | |--------|--------| @@ -211,7 +211,7 @@ SELECT * FROM read_json('records.json', records = false); ``` -
+
| json | |----------------------------------| @@ -232,7 +232,7 @@ SELECT * FROM read_json('arrays.json', records = false); ``` -
+
| json | |-----------| @@ -254,7 +254,7 @@ FROM read_json('todos.json') LIMIT 5; ``` -
+
| userId | id | title | completed | |-------:|---:|-----------------------------------------------------------------|-----------| @@ -275,7 +275,7 @@ CREATE TABLE todos AS DESCRIBE todos; ``` -
+
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|-----|---------|-------| @@ -305,7 +305,7 @@ COPY todos FROM 'todos.json'; SELECT * FROM todos LIMIT 5; ``` -
+
| userId | id | title | completed | |--------|----|-----------------------------------------------------------------|-----------| diff --git a/docs/archive/1.0/data/parquet/metadata.md b/docs/archive/1.0/data/parquet/metadata.md index 9ee5806169b..29917fb6e77 100644 --- a/docs/archive/1.0/data/parquet/metadata.md +++ b/docs/archive/1.0/data/parquet/metadata.md @@ -14,7 +14,7 @@ FROM parquet_metadata('test.parquet'); Below is a table of the columns returned by `parquet_metadata`. -
+
| Field | Type | | ----------------------- | --------------- | @@ -62,7 +62,7 @@ FROM parquet_schema('test.parquet'); Below is a table of the columns returned by `parquet_schema`. -
+
| Field | Type | | --------------- | ------- | @@ -89,7 +89,7 @@ FROM parquet_file_metadata('test.parquet'); Below is a table of the columns returned by `parquet_file_metadata`. -
+
| Field | Type | | ----------------------------| ------- | @@ -112,7 +112,7 @@ FROM parquet_kv_metadata('test.parquet'); Below is a table of the columns returned by `parquet_kv_metadata`. -
+
| Field | Type | | --------- | ------- | diff --git a/docs/archive/1.0/dev/sqllogictest/loops.md b/docs/archive/1.0/dev/sqllogictest/loops.md index 4ca65b36ae3..02672bdf08d 100644 --- a/docs/archive/1.0/dev/sqllogictest/loops.md +++ b/docs/archive/1.0/dev/sqllogictest/loops.md @@ -41,7 +41,7 @@ endloop `foreach` also has a number of preset combinations that should be used when required. In this manner, when new combinations are added to the preset, old tests will automatically pick up these new combinations. -
+
| Preset | Expansion | |----------------|--------------------------------------------------------------| diff --git a/docs/archive/1.0/extensions/json.md b/docs/archive/1.0/extensions/json.md index a8d64dad650..a5af1f242fe 100644 --- a/docs/archive/1.0/extensions/json.md +++ b/docs/archive/1.0/extensions/json.md @@ -241,7 +241,7 @@ Example usage: SELECT * FROM read_json('my_file1.json', columns = {duck: 'INTEGER'}); ``` -
+
| duck | |:---| @@ -257,7 +257,7 @@ FROM read_json( ); ``` -
+
| duck | goose | swan | |:---|:---|:---| @@ -271,7 +271,7 @@ SELECT goose, duck FROM read_json('*.json.gz'); SELECT goose, duck FROM '*.json.gz'; -- equivalent ``` -
+
| goose | duck | |:---|:---| @@ -309,7 +309,7 @@ Can be queried exactly the same as a JSON file that contains `'unstructured'` JS Both can be read as the table: -
+
| duck | goose | |:---|:---| @@ -327,7 +327,7 @@ The `records` parameter specifies whether the JSON contains records that should Results in two columns: -
+
| duck | goose | |:---|:---| @@ -336,7 +336,7 @@ Results in two columns: You can read the same file with `records` set to `'false'`, to get a single column, which is a `STRUCT` containing the data: -
+
| json | |:---| @@ -592,7 +592,7 @@ SELECT FROM example; ``` -
+
| family | species | |------------|------------------------------| @@ -1031,7 +1031,7 @@ FROM ( ); ``` -
+
| (a != b) | (c != d) | (c[0] = d[0]) | (a = c[0]) | (b != c[0]) | |----------|----------|---------------|------------|-------------| diff --git a/docs/archive/1.0/extensions/mysql.md b/docs/archive/1.0/extensions/mysql.md index fd1314626d7..996205d7c37 100644 --- a/docs/archive/1.0/extensions/mysql.md +++ b/docs/archive/1.0/extensions/mysql.md @@ -33,7 +33,7 @@ USE mysqldb; The connection string determines the parameters for how to connect to MySQL as a set of `key=value` pairs. Any options not provided are replaced by their default values, as per the table below. Connection information can also be specified with [environment variables](https://dev.mysql.com/doc/refman/8.3/en/environment-variables.html). If no option is provided explicitly, the MySQL extension tries to read it from an environment variable. -
+
| Setting | Default | Environment variable | |----------|----------------|----------------------| @@ -52,7 +52,7 @@ The tables in the MySQL database can be read as if they were normal DuckDB table SHOW TABLES; ``` -
+
| name | |-----------------| @@ -62,7 +62,7 @@ SHOW TABLES; SELECT * FROM signed_integers; ``` -
+
| t | s | m | i | b | |-----:|-------:|---------:|------------:|---------------------:| diff --git a/docs/archive/1.0/extensions/postgres.md b/docs/archive/1.0/extensions/postgres.md index 362ef1a3aac..e76b902209e 100644 --- a/docs/archive/1.0/extensions/postgres.md +++ b/docs/archive/1.0/extensions/postgres.md @@ -86,7 +86,7 @@ The tables in the PostgreSQL database can be read as if they were normal DuckDB SHOW ALL TABLES; ``` -
+
| name | | ----- | @@ -96,7 +96,7 @@ SHOW ALL TABLES; SELECT * FROM uuids; ``` -
+
| u | | ------------------------------------ | diff --git a/docs/archive/1.0/extensions/spatial.md b/docs/archive/1.0/extensions/spatial.md index 5c8aeaf89cb..4b5d4b53be3 100644 --- a/docs/archive/1.0/extensions/spatial.md +++ b/docs/archive/1.0/extensions/spatial.md @@ -257,7 +257,7 @@ WHERE tags['highway'] != [] LIMIT 5; ``` -
+
| kind | id | tags | refs | lat | lon | ref_roles | ref_types | |------|--------|-------------------------------------------------------------------------------------------------------------------------------------------------------------|------|--------------------|------------|-----------|-----------| diff --git a/docs/archive/1.0/extensions/sqlite.md b/docs/archive/1.0/extensions/sqlite.md index 692487772d0..21ea286f170 100644 --- a/docs/archive/1.0/extensions/sqlite.md +++ b/docs/archive/1.0/extensions/sqlite.md @@ -35,7 +35,7 @@ The tables in the file can be read as if they were normal DuckDB tables, but the SHOW TABLES; ``` -
+
| name | |------------------------| diff --git a/docs/archive/1.0/guides/meta/duckdb_environment.md b/docs/archive/1.0/guides/meta/duckdb_environment.md index 8e60dcc7758..339e1415996 100644 --- a/docs/archive/1.0/guides/meta/duckdb_environment.md +++ b/docs/archive/1.0/guides/meta/duckdb_environment.md @@ -13,7 +13,7 @@ The `version()` function returns the version number of DuckDB. SELECT version() AS version; ``` -
+
| version | |-----------| @@ -25,7 +25,7 @@ Using a `PRAGMA`: PRAGMA version; ``` -
+
| library_version | source_id | |-----------------|------------| diff --git a/docs/archive/1.0/sql/data_types/bitstring.md b/docs/archive/1.0/sql/data_types/bitstring.md index 72c9c48759d..89fa04d4dfc 100644 --- a/docs/archive/1.0/sql/data_types/bitstring.md +++ b/docs/archive/1.0/sql/data_types/bitstring.md @@ -22,7 +22,7 @@ A string encoding a bitstring can be cast to a `BITSTRING`: SELECT '101010'::BITSTRING AS b; ``` -
+
| b | |--------| @@ -44,7 +44,7 @@ Numeric values (integer and float values) can also be converted to a `BITSTRING` SELECT 123::BITSTRING AS b; ``` -
+
| b | |----------------------------------| diff --git a/docs/archive/1.0/sql/data_types/list.md b/docs/archive/1.0/sql/data_types/list.md index 227d20711d4..5957b3301d0 100644 --- a/docs/archive/1.0/sql/data_types/list.md +++ b/docs/archive/1.0/sql/data_types/list.md @@ -53,7 +53,7 @@ Retrieving one or more values from a list can be accomplished using brackets and > This is only needed in our basic examples here, not when working with a list column. > For example, this can't be parsed: `SELECT ['a', 'b', 'c'][1]`. -
+
| Example | Result | |:-----------------------------------------|:-----------| diff --git a/docs/archive/1.0/sql/data_types/numeric.md b/docs/archive/1.0/sql/data_types/numeric.md index 17373b3c533..0ce940244a6 100644 --- a/docs/archive/1.0/sql/data_types/numeric.md +++ b/docs/archive/1.0/sql/data_types/numeric.md @@ -90,7 +90,7 @@ SELECT 0.0 / 0.0 AS x; SELECT -1.0 / 0.0 AS x; ``` -
+
| Expression | DuckDB | IEEE 754 | | :--------- | -----: | --------: | diff --git a/docs/archive/1.0/sql/dialect/postgresql_compatibility.md b/docs/archive/1.0/sql/dialect/postgresql_compatibility.md index 18251bfb83c..662383fcffa 100644 --- a/docs/archive/1.0/sql/dialect/postgresql_compatibility.md +++ b/docs/archive/1.0/sql/dialect/postgresql_compatibility.md @@ -24,7 +24,7 @@ SELECT 'Infinity'::FLOAT - 'Infinity'::FLOAT AS x; SELECT 'Infinity'::FLOAT - 1.0 AS x; ``` -
+
| Expression | DuckDB | PostgreSQL | IEEE 754 | | :---------------------- | -------: | ---------: | --------: | @@ -131,7 +131,7 @@ SELECT * FROM PreservedCase; SELECT table_name FROM duckdb_tables(); ``` -
+
| table_name | | ------------- | @@ -146,7 +146,7 @@ CREATE TABLE MyTaBLe(x INT); SELECT table_name FROM duckdb_tables(); ``` -
+
| table_name | | ---------- | diff --git a/docs/archive/1.0/sql/expressions/subqueries.md b/docs/archive/1.0/sql/expressions/subqueries.md index f73dfdf2bc7..ab93918667b 100644 --- a/docs/archive/1.0/sql/expressions/subqueries.md +++ b/docs/archive/1.0/sql/expressions/subqueries.md @@ -209,7 +209,7 @@ SELECT t FROM (SELECT unnest(generate_series(41, 43)) AS x, 'hello' AS y) t; ``` -
+
| t | |-----------------------| diff --git a/docs/archive/1.0/sql/functions/pattern_matching.md b/docs/archive/1.0/sql/functions/pattern_matching.md index 08f59610e58..49a09a3d94c 100644 --- a/docs/archive/1.0/sql/functions/pattern_matching.md +++ b/docs/archive/1.0/sql/functions/pattern_matching.md @@ -64,7 +64,7 @@ SELECT 'A%c' ILIKE 'a$%c' ESCAPE '$'; -- true There are also alternative characters that can be used as keywords in place of `LIKE` expressions. These enhance PostgreSQL compatibility. -
+
| LIKE-style | PostgreSQL-style | |:---|:---| @@ -93,7 +93,7 @@ SELECT 'abc' NOT SIMILAR TO 'abc'; -- false There are also alternative characters that can be used as keywords in place of `SIMILAR TO` expressions. These follow POSIX syntax. -
+
| `SIMILAR TO`-style | POSIX-style | |:---|:---| @@ -160,7 +160,7 @@ Search the current directory for all files: SELECT * FROM glob('*'); ``` -
+
| file | |---------------| @@ -182,7 +182,7 @@ For example, with a `README.md` file present in the directory, the following que SELECT * FROM glob('README.md'); ``` -
+
| file | |-----------| diff --git a/docs/archive/1.0/sql/query_syntax/filter.md b/docs/archive/1.0/sql/query_syntax/filter.md index e8640436704..9e9964345f0 100644 --- a/docs/archive/1.0/sql/query_syntax/filter.md +++ b/docs/archive/1.0/sql/query_syntax/filter.md @@ -26,7 +26,7 @@ SELECT FROM generate_series(1, 10) tbl(i); ``` -
+
| total_rows | lte_five | odds | |:---|:---|:---| @@ -42,7 +42,7 @@ SELECT FROM generate_series(1, 10) tbl(i); ``` -
+
| lte_five_sum | odds_median | odds_lte_five_median | |:---|:---|:---| @@ -94,7 +94,7 @@ SELECT FROM stacked_data; ``` -
+
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| @@ -114,7 +114,7 @@ SELECT FROM stacked_data; ``` -
+
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| @@ -132,7 +132,7 @@ SELECT FROM stacked_data; ``` -
+
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| diff --git a/docs/data/csv/auto_detection.md b/docs/data/csv/auto_detection.md index 3a82ad38344..4f72c99af77 100644 --- a/docs/data/csv/auto_detection.md +++ b/docs/data/csv/auto_detection.md @@ -96,7 +96,7 @@ After detecting the dialect, the system will attempt to figure out the types of The type detection works by attempting to convert the values in each column to the candidate types. If the conversion is unsuccessful, the candidate type is removed from the set of candidate types for that column. After all samples have been handled – the remaining candidate type with the highest priority is chosen. The default set of candidate types is given below, in order of priority: -
+
| Types | |-----------| @@ -114,7 +114,7 @@ The set of candidate types that should be considered by the CSV reader can be ex In addition to the default set of candidate types, other types that may be specified using the `auto_type_candidates` options are: -
+
| Types | |-----------| @@ -157,7 +157,7 @@ If the ambiguities cannot be resolved by looking at the data the system has a li The system considers the following formats for dates (`dateformat`). Higher entries are chosen over lower entries in case of ambiguities (i.e., ISO 8601 is preferred over `MM-DD-YYYY`). -
+
| dateformat | |------------| @@ -171,7 +171,7 @@ The system considers the following formats for dates (`dateformat`). Higher entr The system considers the following formats for timestamps (`timestampformat`). Higher entries are chosen over lower entries in case of ambiguities. -
+
| timestampformat | |----------------------| diff --git a/docs/data/csv/reading_faulty_csv_files.md b/docs/data/csv/reading_faulty_csv_files.md index 52f4fc29afc..e17158c27c0 100644 --- a/docs/data/csv/reading_faulty_csv_files.md +++ b/docs/data/csv/reading_faulty_csv_files.md @@ -211,7 +211,7 @@ FROM reject_scans; Outputs: -
+
| scan_id | file_id | file_path | delimiter | quote | escape | newline_delimiter | skip_rows | has_header | columns | date_format | timestamp_format | user_arguments | |---------|---------|-----------------------------------|-----------|-------|--------|-------------------|-----------|-----------:|--------------------------------------|-------------|------------------|--------------------| @@ -223,7 +223,7 @@ FROM reject_errors; Outputs: -
+
| scan_id | file_id | line | line_byte_position | byte_position | column_idx | column_name | error_type | csv_line | error_message | |---------|---------|------|--------------------|---------------|------------|-------------|------------|---------------------|------------------------------------------------------------------------------------| diff --git a/docs/data/json/caveats.md b/docs/data/json/caveats.md index 262feb7c1b4..84bbdde3588 100644 --- a/docs/data/json/caveats.md +++ b/docs/data/json/caveats.md @@ -25,7 +25,7 @@ FROM ( ); ``` -
+
| (a != b) | (c != d) | (c[0] = d[0]) | (a = c[0]) | (b != c[0]) | |----------|----------|---------------|------------|-------------| diff --git a/docs/data/json/format_settings.md b/docs/data/json/format_settings.md index af3c0b59f19..8f1da600c36 100644 --- a/docs/data/json/format_settings.md +++ b/docs/data/json/format_settings.md @@ -32,7 +32,7 @@ SELECT * FROM read_json('records.json', format = 'newline_delimited'); ``` -
+
| key1 | key2 | |--------|--------| @@ -58,7 +58,7 @@ SELECT * FROM read_json('records-in-array.json', format = 'array'); ``` -
+
| key1 | key2 | |--------|--------| @@ -91,7 +91,7 @@ SELECT * FROM read_json('unstructured.json', format = 'unstructured'); ``` -
+
| key1 | key2 | |--------|--------| @@ -117,7 +117,7 @@ SELECT * FROM read_json('records.json', records = true); ``` -
+
| key1 | key2 | |--------|--------| @@ -132,7 +132,7 @@ SELECT * FROM read_json('records.json', records = false); ``` -
+
| json | |----------------------------------| @@ -153,7 +153,7 @@ SELECT * FROM read_json('arrays.json', records = false); ``` -
+
| json | |-----------| diff --git a/docs/data/json/json_functions.md b/docs/data/json/json_functions.md index a15f262ef67..28411567068 100644 --- a/docs/data/json/json_functions.md +++ b/docs/data/json/json_functions.md @@ -142,7 +142,7 @@ SELECT FROM example; ``` -
+
| family | species | |------------|------------------------------| diff --git a/docs/data/json/loading_json.md b/docs/data/json/loading_json.md index ead143744b8..e5fc7528a98 100644 --- a/docs/data/json/loading_json.md +++ b/docs/data/json/loading_json.md @@ -122,7 +122,7 @@ Example usage: SELECT * FROM read_json('my_file1.json', columns = {duck: 'INTEGER'}); ``` -
+
| duck | |:---| @@ -138,7 +138,7 @@ FROM read_json( ); ``` -
+
| duck | goose | swan | |:---|:---|:---| @@ -152,7 +152,7 @@ SELECT goose, duck FROM read_json('*.json.gz'); SELECT goose, duck FROM '*.json.gz'; -- equivalent ``` -
+
| goose | duck | |:---|:---| @@ -190,7 +190,7 @@ Can be queried exactly the same as a JSON file that contains `'unstructured'` JS Both can be read as the table: -
+
| duck | goose | |:---|:---| @@ -208,7 +208,7 @@ The `records` parameter specifies whether the JSON contains records that should Results in two columns: -
+
| duck | goose | |:---|:---| @@ -217,7 +217,7 @@ Results in two columns: You can read the same file with `records` set to `'false'`, to get a single column, which is a `STRUCT` containing the data: -
+
| json | |:---| @@ -333,7 +333,7 @@ CREATE TABLE todos AS DESCRIBE todos; ``` -
+
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|-----|---------|-------| diff --git a/docs/data/parquet/metadata.md b/docs/data/parquet/metadata.md index 9966dbe3879..80ef59a26c5 100644 --- a/docs/data/parquet/metadata.md +++ b/docs/data/parquet/metadata.md @@ -14,7 +14,7 @@ FROM parquet_metadata('test.parquet'); Below is a table of the columns returned by `parquet_metadata`. -
+
| Field | Type | | ----------------------- | --------------- | @@ -62,7 +62,7 @@ FROM parquet_schema('test.parquet'); Below is a table of the columns returned by `parquet_schema`. -
+
| Field | Type | | --------------- | ------- | @@ -89,7 +89,7 @@ FROM parquet_file_metadata('test.parquet'); Below is a table of the columns returned by `parquet_file_metadata`. -
+
| Field | Type | | ----------------------------| ------- | @@ -112,7 +112,7 @@ FROM parquet_kv_metadata('test.parquet'); Below is a table of the columns returned by `parquet_kv_metadata`. -
+
| Field | Type | | --------- | ------- | diff --git a/docs/dev/sqllogictest/loops.md b/docs/dev/sqllogictest/loops.md index ddb8a692e79..2d690f24ccf 100644 --- a/docs/dev/sqllogictest/loops.md +++ b/docs/dev/sqllogictest/loops.md @@ -41,7 +41,7 @@ endloop `foreach` also has a number of preset combinations that should be used when required. In this manner, when new combinations are added to the preset, old tests will automatically pick up these new combinations. -
+
| Preset | Expansion | |----------------|--------------------------------------------------------------| diff --git a/docs/extensions/aws.md b/docs/extensions/aws.md index a4704511a95..6670aa25153 100644 --- a/docs/extensions/aws.md +++ b/docs/extensions/aws.md @@ -49,7 +49,7 @@ To load the AWS credentials, run: CALL load_aws_credentials(); ``` -
+
| loaded_access_key_id | loaded_secret_access_key | loaded_session_token | loaded_region | |----------------------|--------------------------|----------------------|---------------| @@ -61,7 +61,7 @@ The function takes a string parameter to specify a specific profile: CALL load_aws_credentials('minio-testing-2'); ``` -
+
| loaded_access_key_id | loaded_secret_access_key | loaded_session_token | loaded_region | |----------------------|--------------------------|----------------------|---------------| @@ -73,7 +73,7 @@ There are several parameters to tweak the behavior of the call: CALL load_aws_credentials('minio-testing-2', set_region = false, redact_secret = false); ``` -
+
| loaded_access_key_id | loaded_secret_access_key | loaded_session_token | loaded_region | |----------------------|------------------------------|----------------------|---------------| diff --git a/docs/extensions/iceberg.md b/docs/extensions/iceberg.md index 84bb62b6619..d618221d0c6 100644 --- a/docs/extensions/iceberg.md +++ b/docs/extensions/iceberg.md @@ -73,7 +73,7 @@ SELECT * FROM iceberg_metadata('data/iceberg/lineitem_iceberg', allow_moved_paths = true); ``` -
+
| manifest_path | manifest_sequence_number | manifest_content | status | content | file_path | file_format | record_count | |------------------------------------------------------------------------|--------------------------|------------------|---------|----------|------------------------------------------------------------------------------------|-------------|--------------| @@ -87,7 +87,7 @@ SELECT * FROM iceberg_snapshots('data/iceberg/lineitem_iceberg'); ``` -
+
| sequence_number | snapshot_id | timestamp_ms | manifest_list | |-----------------|---------------------|-------------------------|------------------------------------------------------------------------------------------------| diff --git a/docs/extensions/mysql.md b/docs/extensions/mysql.md index 4e6993578d4..84d4850a28c 100644 --- a/docs/extensions/mysql.md +++ b/docs/extensions/mysql.md @@ -33,7 +33,7 @@ USE mysqldb; The connection string determines the parameters for how to connect to MySQL as a set of `key=value` pairs. Any options not provided are replaced by their default values, as per the table below. Connection information can also be specified with [environment variables](https://dev.mysql.com/doc/refman/8.3/en/environment-variables.html). If no option is provided explicitly, the MySQL extension tries to read it from an environment variable. -
+
| Setting | Default | Environment variable | |-------------|----------------|----------------------| @@ -126,7 +126,7 @@ The tables in the MySQL database can be read as if they were normal DuckDB table SHOW ALL TABLES; ``` -
+
| name | |-----------------| @@ -136,7 +136,7 @@ SHOW ALL TABLES; SELECT * FROM signed_integers; ``` -
+
| t | s | m | i | b | |-----:|-------:|---------:|------------:|---------------------:| diff --git a/docs/extensions/postgres.md b/docs/extensions/postgres.md index ff6e41b9b1a..50c124c55b5 100644 --- a/docs/extensions/postgres.md +++ b/docs/extensions/postgres.md @@ -142,7 +142,7 @@ The tables in the PostgreSQL database can be read as if they were normal DuckDB SHOW ALL TABLES; ``` -
+
| name | | ----- | @@ -152,7 +152,7 @@ SHOW ALL TABLES; SELECT * FROM uuids; ``` -
+
| u | | ------------------------------------ | diff --git a/docs/extensions/sqlite.md b/docs/extensions/sqlite.md index 753e799a604..6911f098437 100644 --- a/docs/extensions/sqlite.md +++ b/docs/extensions/sqlite.md @@ -35,7 +35,7 @@ The tables in the file can be read as if they were normal DuckDB tables, but the SHOW TABLES; ``` -
+
| name | |------------------------| diff --git a/docs/extensions/working_with_extensions.md b/docs/extensions/working_with_extensions.md index 36bf966d580..00cd420f4cd 100644 --- a/docs/extensions/working_with_extensions.md +++ b/docs/extensions/working_with_extensions.md @@ -111,7 +111,7 @@ SELECT extension_name, extension_version, installed_from, install_mode FROM duck This outputs: -
+
| extensions_name | extensions_version | installed_from | install_mode | |:----------------|:-------------------|:---------------|:-------------| diff --git a/docs/guides/meta/duckdb_environment.md b/docs/guides/meta/duckdb_environment.md index c2211c2ca58..bd187787bd8 100644 --- a/docs/guides/meta/duckdb_environment.md +++ b/docs/guides/meta/duckdb_environment.md @@ -13,7 +13,7 @@ The `version()` function returns the version number of DuckDB. SELECT version() AS version; ``` -
+
| version | |-----------| @@ -25,7 +25,7 @@ Using a `PRAGMA`: PRAGMA version; ``` -
+
| library_version | source_id | |-----------------|------------| diff --git a/docs/guides/sql_features/full_text_search.md b/docs/guides/sql_features/full_text_search.md index 985045e8295..f05b17b1d70 100644 --- a/docs/guides/sql_features/full_text_search.md +++ b/docs/guides/sql_features/full_text_search.md @@ -19,7 +19,7 @@ CREATE TABLE corpus AS DESCRIBE corpus; ``` -
+
| column_name | column_type | null | key | default | extra | |-------------|-------------|------|------|---------|-------| diff --git a/docs/sql/data_types/bitstring.md b/docs/sql/data_types/bitstring.md index 280225d74c7..a279575bc2d 100644 --- a/docs/sql/data_types/bitstring.md +++ b/docs/sql/data_types/bitstring.md @@ -22,7 +22,7 @@ A string encoding a bitstring can be cast to a `BITSTRING`: SELECT '101010'::BITSTRING AS b; ``` -
+
| b | |--------| @@ -44,7 +44,7 @@ Numeric values (integer and float values) can also be converted to a `BITSTRING` SELECT 123::BITSTRING AS b; ``` -
+
| b | |----------------------------------| diff --git a/docs/sql/data_types/list.md b/docs/sql/data_types/list.md index a5942158a0f..5052d2d63fd 100644 --- a/docs/sql/data_types/list.md +++ b/docs/sql/data_types/list.md @@ -53,7 +53,7 @@ Retrieving one or more values from a list can be accomplished using brackets and > This is only needed in our basic examples here, not when working with a list column. > For example, this can't be parsed: `SELECT ['a', 'b', 'c'][1]`. -
+
| Example | Result | |:-----------------------------------------|:-----------| diff --git a/docs/sql/data_types/numeric.md b/docs/sql/data_types/numeric.md index 80a6fff80c5..c50366a561c 100644 --- a/docs/sql/data_types/numeric.md +++ b/docs/sql/data_types/numeric.md @@ -86,7 +86,7 @@ SELECT 'nan' > sqrt(2) ``` -
+
| (sqrt(2) > '-inf') | ('nan' > sqrt(2)) | |-------------------:|------------------:| diff --git a/docs/sql/dialect/postgresql_compatibility.md b/docs/sql/dialect/postgresql_compatibility.md index c3022d305af..9408f324715 100644 --- a/docs/sql/dialect/postgresql_compatibility.md +++ b/docs/sql/dialect/postgresql_compatibility.md @@ -22,7 +22,7 @@ SELECT 'Infinity'::FLOAT - 'Infinity'::FLOAT AS x; SELECT 'Infinity'::FLOAT - 1.0 AS x; ``` -
+
| Expression | PostgreSQL | DuckDB | IEEE 754 | | :---------------------- | ---------: | --------: | --------: | @@ -131,7 +131,7 @@ SELECT * FROM PreservedCase; SELECT table_name FROM duckdb_tables(); ``` -
+
| table_name | | ------------- | @@ -146,7 +146,7 @@ CREATE TABLE MyTaBLe(x INT); SELECT table_name FROM duckdb_tables(); ``` -
+
| table_name | | ---------- | @@ -164,7 +164,7 @@ SELECT 1 == 1 AS t; DuckDB returns: -
+
| t | | ---: | @@ -226,7 +226,7 @@ DuckDB runs the statement and creates the table successfully, confirmed by the f DESCRIBE myschema.mytable; ``` -
+
| column_name | column_type | null | key | default | extra | |-------------|------------------|------|------|---------|-------| diff --git a/docs/sql/expressions/subqueries.md b/docs/sql/expressions/subqueries.md index 2db88e073f5..9d106f6154e 100644 --- a/docs/sql/expressions/subqueries.md +++ b/docs/sql/expressions/subqueries.md @@ -209,7 +209,7 @@ SELECT t FROM (SELECT unnest(generate_series(41, 43)) AS x, 'hello' AS y) t; ``` -
+
| t | |-----------------------| diff --git a/docs/sql/functions/list.md b/docs/sql/functions/list.md index 2788f7153ee..b202c39d995 100644 --- a/docs/sql/functions/list.md +++ b/docs/sql/functions/list.md @@ -426,7 +426,7 @@ SELECT [lower(x) FOR x IN strings] AS strings FROM (VALUES (['Hello', '', 'World'])) t(strings); ``` -
+
| strings | |------------------| @@ -437,7 +437,7 @@ SELECT [upper(x) FOR x IN strings IF len(x) > 0] AS strings FROM (VALUES (['Hello', '', 'World'])) t(strings); ``` -
+
| strings | |----------------| @@ -450,7 +450,7 @@ In the following example, we use `x, i`, where `x` is the value and `i` is the p SELECT [4, 5, 6] AS l, [x FOR x, i IN l IF i != 2] filtered; ``` -
+
| l | filtered | |-----------|----------| diff --git a/docs/sql/functions/pattern_matching.md b/docs/sql/functions/pattern_matching.md index a76594b2ecd..8869b3f653d 100644 --- a/docs/sql/functions/pattern_matching.md +++ b/docs/sql/functions/pattern_matching.md @@ -64,7 +64,7 @@ SELECT 'A%c' ILIKE 'a$%c' ESCAPE '$'; -- true There are also alternative characters that can be used as keywords in place of `LIKE` expressions. These enhance PostgreSQL compatibility. -
+
| LIKE-style | PostgreSQL-style | |:---|:---| @@ -93,7 +93,7 @@ SELECT 'abc' NOT SIMILAR TO 'abc'; -- false There are also alternative characters that can be used as keywords in place of `SIMILAR TO` expressions. These follow POSIX syntax. -
+
| `SIMILAR TO`-style | POSIX-style | |:---|:---| @@ -160,7 +160,7 @@ Search the current directory for all files: SELECT * FROM glob('*'); ``` -
+
| file | |---------------| @@ -182,7 +182,7 @@ For example, with a `README.md` file present in the directory, the following que SELECT * FROM glob('README.md'); ``` -
+
| file | |-----------| diff --git a/docs/sql/query_syntax/filter.md b/docs/sql/query_syntax/filter.md index 21a0d2fcd73..11bf8ee742e 100644 --- a/docs/sql/query_syntax/filter.md +++ b/docs/sql/query_syntax/filter.md @@ -26,7 +26,7 @@ SELECT FROM generate_series(1, 10) tbl(i); ``` -
+
| total_rows | lte_five | odds | |:---|:---|:---| @@ -42,7 +42,7 @@ SELECT FROM generate_series(1, 10) tbl(i); ``` -
+
| lte_five_sum | odds_median | odds_lte_five_median | |:---|:---|:---| @@ -94,7 +94,7 @@ SELECT FROM stacked_data; ``` -
+
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| @@ -114,7 +114,7 @@ SELECT FROM stacked_data; ``` -
+
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| @@ -132,7 +132,7 @@ SELECT FROM stacked_data; ``` -
+
| 2022 | 2023 | 2024 | 2025 | NULLs | |:---|:---|:---|:---|:---| From 8cbf841e07ab8328b0a776274c98ad73ad40bfd0 Mon Sep 17 00:00:00 2001 From: Gabor Szarnyas Date: Sun, 8 Dec 2024 21:57:54 +0100 Subject: [PATCH 4/4] CSS: Remove narrow_table --- css/docu.scss | 13 ++----------- 1 file changed, 2 insertions(+), 11 deletions(-) diff --git a/css/docu.scss b/css/docu.scss index 23c0384edeb..2f9a394b30f 100644 --- a/css/docu.scss +++ b/css/docu.scss @@ -924,13 +924,6 @@ table thead th { .media_table + table td:nth-child(4) { width: 80px; } -.narrow_table + table{ - width: auto; - min-width: auto; - td{ - min-width: 40px; - } -} .center_aligned_header_table + table thead th { text-align: center !important; } @@ -1013,7 +1006,6 @@ table{ h4 + table, -h4 + .narrow_table, h4 + .highlighter-rouge{ margin-top: 1em; } @@ -1021,8 +1013,7 @@ form { background: #f2f2f2; padding: 20px; } -.highlighter-rouge + table, -.highlighter-rouge + .narrow_table{ +.highlighter-rouge + table { margin-top: 25px; } .outer { @@ -1478,4 +1469,4 @@ body.documentation.installation{ max-width: 320px; font-size: 14px; } -} \ No newline at end of file +}