Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions website/docs/engine-flink/ddl.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@ The following properties can be set if using the Fluss catalog:
| bootstrap.servers | required | (none) | Comma separated list of Fluss servers. |
| default-database | optional | fluss | The default database to use when switching to this catalog. |
| client.security.protocol | optional | PLAINTEXT | The security protocol used to communicate with brokers. Currently, only `PLAINTEXT` and `SASL` are supported, the configuration value is case insensitive. |
| `client.security.{protocol}.*` | optional | (none) | Client-side configuration properties for a specific authentication protocol. E.g., client.security.sasl.jaas.config. More Details in [authentication](../security/authentication.md) |
| `client.security.{protocol}.*` | optional | (none) | Client-side configuration properties for a specific authentication protocol. E.g., client.security.sasl.jaas.config. More Details in [authentication](security/authentication.md) |
| `{lake-format}.*` | optional | (none) | Extra properties to be passed to the lake catalog. This is useful for configuring sensitive settings, such as the username and password required for lake catalog authentication. E.g., `paimon.jdbc.password = pass`. |

The following statements assume that the current catalog has been switched to the Fluss catalog using the `USE CATALOG <catalog_name>` statement.
Expand Down Expand Up @@ -62,7 +62,7 @@ DROP DATABASE my_db;

### Primary Key Table

The following SQL statement will create a [Primary Key Table](table-design/table-types/pk-table/index.md) with a primary key consisting of shop_id and user_id.
The following SQL statement will create a [Primary Key Table](table-design/table-types/pk-table.md) with a primary key consisting of shop_id and user_id.
```sql title="Flink SQL"
CREATE TABLE my_pk_table (
shop_id BIGINT,
Expand Down
2 changes: 1 addition & 1 deletion website/docs/engine-flink/delta-joins.md
Original file line number Diff line number Diff line change
Expand Up @@ -130,7 +130,7 @@ For example:
- Full primary key: `(city_id, order_id)`
- Bucket key: `city_id`

This yields an **index** on the prefix key `city_id`, so that you can perform [Prefix Key Lookup](/docs/engine-flink/lookups/#prefix-lookup) by the `city_id`.
This yields an **index** on the prefix key `city_id`, so that you can perform [Prefix Key Lookup](/engine-flink/lookups.md#prefix-lookup) by the `city_id`.

In this setup:
* The delta join operator uses the prefix key (`city_id`) to retrieve only relevant right-side records matching each left-side event.
Expand Down
4 changes: 2 additions & 2 deletions website/docs/engine-flink/options.md
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ See more details about [ALTER TABLE ... SET](engine-flink/ddl.md#set-properties)
| table.datalake.freshness | Duration | 3min | It defines the maximum amount of time that the datalake table's content should lag behind updates to the Fluss table. Based on this target freshness, the Fluss service automatically moves data from the Fluss table and updates to the datalake table, so that the data in the datalake table is kept up to date within this target. If the data does not need to be as fresh, you can specify a longer target freshness time to reduce costs. |
| table.datalake.auto-compaction | Boolean | false | If true, compaction will be triggered automatically when tiering service writes to the datalake. It is disabled by default. |
| table.datalake.auto-expire-snapshot | Boolean | false | If true, snapshot expiration will be triggered automatically when tiering service commits to the datalake. It is disabled by default. |
| table.merge-engine | Enum | (None) | Defines the merge engine for the primary key table. By default, primary key table uses the [default merge engine(last_row)](table-design/table-types/pk-table/merge-engines/default.md). It also supports two merge engines are `first_row`, `versioned` and `aggregation`. The [first_row merge engine](table-design/table-types/pk-table/merge-engines/first-row.md) will keep the first row of the same primary key. The [versioned merge engine](table-design/table-types/pk-table/merge-engines/versioned.md) will keep the row with the largest version of the same primary key. The `aggregation` merge engine will aggregate rows with the same primary key using field-level aggregate functions. |
| table.merge-engine | Enum | (None) | Defines the merge engine for the primary key table. By default, primary key table uses the [default merge engine(last_row)](table-design/merge-engines/default.md). It also supports two merge engines are `first_row`, `versioned` and `aggregation`. The [first_row merge engine](table-design/merge-engines/first-row.md) will keep the first row of the same primary key. The [versioned merge engine](table-design/merge-engines/versioned.md) will keep the row with the largest version of the same primary key. The `aggregation` merge engine will aggregate rows with the same primary key using field-level aggregate functions. |
| table.merge-engine.versioned.ver-column | String | (None) | The column name of the version column for the `versioned` merge engine. If the merge engine is set to `versioned`, the version column must be set. |
| table.delete.behavior | Enum | ALLOW | Controls the behavior of delete operations on primary key tables. Three modes are supported: `ALLOW` (default for default merge engine) - allows normal delete operations; `IGNORE` - silently ignores delete requests without errors; `DISABLE` - rejects delete requests and throws explicit errors. This configuration provides system-level guarantees for some downstream pipelines (e.g., Flink Delta Join) that must not receive any delete events in the changelog of the table. For tables with `first_row` or `versioned` or `aggregation` merge engines, this option is automatically set to `IGNORE` and cannot be overridden. Note: For `aggregation` merge engine, when set to `allow`, delete operations will remove the entire record. This configuration only applicable to primary key tables. |
| table.changelog.image | Enum | FULL | Defines the changelog image mode for primary key tables. This configuration is inspired by similar settings in database systems like MySQL's `binlog_row_image` and PostgreSQL's `replica identity`. Two modes are supported: `FULL` (default) - produces both UPDATE_BEFORE and UPDATE_AFTER records for update operations, capturing complete information about updates and allowing tracking of previous values; `WAL` - does not produce UPDATE_BEFORE records. Only INSERT, UPDATE_AFTER (and DELETE if allowed) records are emitted. When WAL mode is enabled with default merge engine (no merge engine configured) and full row updates (not partial update), an optimization is applied to skip looking up old values, and in this case INSERT operations are converted to UPDATE_AFTER events. This mode reduces storage and transmission costs but loses the ability to track previous values. Only applicable to primary key tables. |
Expand Down Expand Up @@ -157,4 +157,4 @@ See more details about [ALTER TABLE ... SET](engine-flink/ddl.md#set-properties)
| client.filesystem.security.token.renewal.time-ratio | Double | 0.75 | Ratio of the token's expiration time when new credentials for access filesystem should be re-obtained. |
| client.metrics.enabled | Boolean | false | Enable metrics for client. When metrics is enabled, the client will collect metrics and report by the JMX metrics reporter. |
| client.security.protocol | String | PLAINTEXT | The security protocol used to communicate with brokers. Currently, only `PLAINTEXT` and `SASL` are supported, the configuration value is case insensitive. |
| client.security.\{protocol\}.* | optional | (none) | Client-side configuration properties for a specific authentication protocol. E.g., client.security.sasl.jaas.config. More Details in [authentication](../security/authentication.md) |
| client.security.\{protocol\}.* | optional | (none) | Client-side configuration properties for a specific authentication protocol. E.g., client.security.sasl.jaas.config. More Details in [authentication](security/authentication.md) |
2 changes: 1 addition & 1 deletion website/docs/engine-flink/procedures.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ SHOW PROCEDURES;

## Access Control Procedures

Fluss provides procedures to manage Access Control Lists (ACLs) for security and authorization. See the [Security](../security/overview.md) documentation for more details.
Fluss provides procedures to manage Access Control Lists (ACLs) for security and authorization. See the [Security](/security/overview.md) documentation for more details.

### add_acl

Expand Down
4 changes: 2 additions & 2 deletions website/docs/maintenance/operations/graceful-shutdown.md
Original file line number Diff line number Diff line change
Expand Up @@ -131,6 +131,6 @@ Monitor shutdown-related metrics:

## See Also

- [Configuration](../configuration.md)
- [Monitoring and Observability](../observability/monitor-metrics.md)
- [Configuration](maintenance/configuration.md)
- [Monitoring and Observability](maintenance/observability/monitor-metrics.md)
- [Upgrading Fluss](upgrading.md)
2 changes: 1 addition & 1 deletion website/docs/maintenance/operations/updating-configs.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ Currently, the supported dynamically updatable server configurations include:
- `kv.rocksdb.shared-rate-limiter.bytes-per-sec`: Control RocksDB flush and compaction write rate shared across all RocksDB instances on the TabletServer. The rate limiter is always enabled. Set to a lower value (e.g., 100MB) to limit the rate, or a very high value to effectively disable rate limiting.


You can update the configuration of a cluster with [Java client](#using-java-client) or [Flink Procedures](../../engine-flink/procedures.md#cluster-configuration-procedures).
You can update the configuration of a cluster with [Java client](#using-java-client) or [Flink Procedures](engine-flink/procedures.md#cluster-configuration-procedures).

### Using Java Client

Expand Down
2 changes: 1 addition & 1 deletion website/docs/streaming-lakehouse/overview.md
Original file line number Diff line number Diff line change
Expand Up @@ -43,4 +43,4 @@ Some powerful features it provides are:
- **Analytical Streams**: The union reads help data streams to have the powerful analytics capabilities. This reduces complexity when developing streaming applications, simplifies debugging, and allows for immediate access to live data insights.
- **Connect to Lakehouse Ecosystem**: Fluss keeps the table metadata in sync with data lake catalogs while compacting data into Lakehouse. As a result, external engines like Spark, StarRocks, Flink, and Trino can read the data directly. They simply connect to the data lake catalog.

Currently, Fluss supports [Paimon](integrate-data-lakes/paimon.md), [Iceberg](integrate-data-lakes/iceberg.md), and [Lance](integrate-data-lakes/lance.md) as Lakehouse Storage, more kinds of data lake formats are on the roadmap.
Currently, Fluss supports [Paimon](streaming-lakehouse/integrate-data-lakes/paimon.md), [Iceberg](streaming-lakehouse/integrate-data-lakes/iceberg.md), and [Lance](streaming-lakehouse/integrate-data-lakes/lance.md) as Lakehouse Storage, more kinds of data lake formats are on the roadmap.
2 changes: 1 addition & 1 deletion website/docs/table-design/data-types.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
---
title: "Data Types"
sidebar_position: 10
sidebar_position: 5
---

# Data Types
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
{
"label": "Merge Engines",
"position": 2
"position": 3
}
Original file line number Diff line number Diff line change
Expand Up @@ -1012,8 +1012,8 @@ For detailed information about Exactly-Once implementation, please refer to: [FI

## See Also

- [Default Merge Engine](./default.md)
- [FirstRow Merge Engine](./first-row.md)
- [Versioned Merge Engine](./versioned.md)
- [Primary Key Tables](../index.md)
- [Fluss Client API](../../../../apis/java-client.md)
- [Default Merge Engine](table-design/merge-engines/default.md)
- [FirstRow Merge Engine](table-design/merge-engines/first-row.md)
- [Versioned Merge Engine](table-design/merge-engines/versioned.md)
- [Primary Key Tables](table-design/table-types/pk-table.md)
- [Fluss Client API](apis/java-client.md)
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ sidebar_position: 2
## Overview

The **Default Merge Engine** behaves as a LastRow merge engine that retains the latest record for a given primary key. It supports all the operations: `INSERT`, `UPDATE`, `DELETE`.
Additionally, the default merge engine supports [Partial Update](table-design/table-types/pk-table/index.md#partial-update), which preserves the latest values for the specified update columns.
Additionally, the default merge engine supports [Partial Update](table-design/table-types/pk-table.md#partial-update), which preserves the latest values for the specified update columns.
If the `'table.merge-engine'` property is not explicitly defined in the table properties when creating a Primary Key Table, the default merge engine will be applied automatically.


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ However, users can specify a different merge engine to customize the merging beh

The following merge engines are supported:

1. [Default Merge Engine (LastRow)](table-design/table-types/pk-table/merge-engines/default.md)
2. [FirstRow Merge Engine](table-design/table-types/pk-table/merge-engines/first-row.md)
3. [Versioned Merge Engine](table-design/table-types/pk-table/merge-engines/versioned.md)
4. [Aggregation Merge Engine](table-design/table-types/pk-table/merge-engines/aggregation.md)
1. [Default Merge Engine (LastRow)](default.md)
2. [FirstRow Merge Engine](first-row.md)
3. [Versioned Merge Engine](versioned.md)
4. [Aggregation Merge Engine](aggregation.md)
4 changes: 2 additions & 2 deletions website/docs/table-design/overview.md
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
---
sidebar_label: Overview
title: Table Overview
sidebar_position: 2
sidebar_position: 1
---

# Table Overview
Expand All @@ -20,7 +20,7 @@ Tables are classified into two types based on the presence of a primary key:
- Used for updating and managing data in business databases.
- Support INSERT, UPDATE, and DELETE operations based on the defined primary key.

A Table becomes a [Partitioned Table](data-distribution/partitioning.md) when a partition column is defined. Data with the same partition value is stored in the same partition. Partition columns can be applied to both Log Tables and Primary Key Tables, but with specific considerations:
A Table becomes a [Partitioned Table](/table-design/data-distribution/partitioning.md) when a partition column is defined. Data with the same partition value is stored in the same partition. Partition columns can be applied to both Log Tables and Primary Key Tables, but with specific considerations:
- **For Log Tables**, partitioning is commonly used for log data, typically based on date columns, to facilitate data separation and cleaning.
- **For Primary Key Tables**, the partition column must be a subset of the primary key to ensure uniqueness.

Expand Down
Loading