Problem Description: Data quality error during loading.
Solution:
Show Load command.Problem Description: Error -235 occurs when the load frequency is too high and data hasn't been compacted in time, exceeding version limits.
Solution:
max_tablet_version_num parameter in be.conf, it is recommended not to exceed 5000.Problem Description: Error -238 occurs when the number of segments under a single rowset exceeds the limit.
Common Causes:
Problem Description: Data load is successful but temporarily not visible.
Cause: Usually due to transaction publish delay caused by system resource pressure.
Problem Description: In shared-nothing mode, too many partitions and tablets are involved in a single load, exceeding the transaction size limit.
Solution:
Problem Description: Usually caused by Windows line endings.
Solution: Specify the correct line delimiter: -H "line_delimiter:\r\n"
Problem Description: CSV data with quotes becomes null after import.
Solution: Use the trim_double_quotes parameter to remove double quotes around fields.
When column names contain special characters, use single quotes with backticks to specify the columns parameter:
curl --location-trusted -u root:"" \ -H 'columns:`@coltime`,colint,colvar' \ -T a.csv \ -H "column_separator:," \ http://127.0.0.1:8030/api/db/loadtest/_stream_load
| Issue Description | Trigger Conditions | Impact Scope | Temporary Solution | Affected Versions | Fixed Versions | Fix PR |
|---|---|---|---|---|---|---|
| When at least one job times out while connecting to Kafka, it affects the import of other jobs, slowing down global Routine Load imports. | At least one job times out while connecting to Kafka. | Shared-nothing and shared-storage | Stop or manually pause the job to resolve the issue. | <2.1.9 <3.0.5 | 2.1.9 3.0.5 | #47530 |
| User data may be lost after restarting the FE Master. | The job's offset is set to OFFSET_END, and the FE is restarted. | Shared-storage | Change the consumption mode to OFFSET_BEGINNING. | 3.0.2-3.0.4 | 3.0.5 | #46149 |
| A large number of small transactions are generated during import, causing compaction to fail and resulting in continuous -235 errors. | Doris consumes data too quickly, or Kafka data flow is in small batches. | Shared-nothing and shared-storage | Pause the Routine Load job and execute the following command: ALTER ROUTINE LOAD FOR jobname FROM kafka ("property.enable.partition.eof" = "false"); | <2.1.8 <3.0.4 | 2.1.8 3.0.4 | #45528, #44949, #39975 |
| Kafka third-party library destructor hangs, causing data consumption to fail. | Kafka topic deletion (possibly other conditions). | Shared-nothing and shared-storage | Restart all BE nodes. | <2.1.8 <3.0.4 | 2.1.8 3.0.4 | #44913 |
| Routine Load scheduling hangs. | Timeout occurs when FE aborts a transaction in Meta Service. | Shared-storage | Restart the FE node. | <3.0.2 | 3.0.2 | #41267 |
| Routine Load restart issue. | Restarting BE nodes. | Shared-nothing and shared-storage | Manually resume the job. | <2.1.7 <3.0.2 | 2.1.7 3.0.2 | #3727 |
| Optimization Content | Applied Versions | Corresponding PR |
|---|---|---|
| Increased the timeout duration for Routine Load. | 2.1.7 3.0.3 | #42042, #40818 |
Adjusted the default value of max_batch_interval. | 2.1.8 3.0.3 | #42491 |
Removed the restriction on max_batch_interval. | 2.1.5 3.0.0 | #29071 |
Adjusted the default values of max_batch_rows and max_batch_size. | 2.1.5 3.0.0 | #36632 |
| Optimization Content | Applied Versions | Corresponding PR |
|---|---|---|
| Added observability-related metrics. | 3.0.5 | #48209, #48171, #48963 |
Problem Description: Routine Load cannot get the latest Kafka offset.
Common Causes:
Problem Description: Routine Load cannot get Kafka Topic Partition Meta.
Common Causes:
Problem Description: The consumed offset doesn't exist in Kafka, possibly because it has been cleaned up by Kafka.
Solution: