Skip to content

Commit 07cae40

Browse files
authored
[docs] Fix typo in website/docs/streaming-lakehouse/overview.md (#260)
1 parent 2435149 commit 07cae40

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

website/docs/streaming-lakehouse/overview.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ sidebar_position: 1
88

99
Lakehouse represents a new, open architecture that combines the best elements of data lakes and data warehouses.
1010
It combines data lake scalability and cost-effectiveness with data warehouse reliability and performance.
11-
The well known data lake format such like [Apache Iceberg](https://iceberg.apache.org/), [Apache Paimon](https://paimon.apache.org/), [Apache Hudi](https://hudi.apache.org/) and [Delta Lake](https://delta.io/) play key roles in the Lakehouse architecture,
11+
The well-known data lake format such like [Apache Iceberg](https://iceberg.apache.org/), [Apache Paimon](https://paimon.apache.org/), [Apache Hudi](https://hudi.apache.org/) and [Delta Lake](https://delta.io/) play key roles in the Lakehouse architecture,
1212
facilitating a harmonious balance between data storage, reliability, and analytical capabilities within a single, unified platform.
1313

1414
Lakehouse, as a modern architecture, is effective in addressing the complex needs of data management and analytics.
@@ -17,7 +17,7 @@ With these data lake formats, you will get into a contradictory situation:
1717

1818
1. If you require low latency, then you write and commit frequently, which means many small Parquet files. This becomes inefficient for
1919
reads which must now deal with masses of small files.
20-
2. If you require read efficiency, then you accumulate data until you can write to large Parquet files, but this introduces
20+
2. If you require reading efficiency, then you accumulate data until you can write to large Parquet files, but this introduces
2121
much higher latency.
2222

2323
Overall, these data lake formats typically achieve data freshness at best within minute-level granularity, even under optimal usage conditions.

0 commit comments

Comments
 (0)