import ChangeLog from ‘../changelog/connector-amazonsqs.md’;
AmazonSqs source connector
Spark
Flink
SeaTunnel Zeta
Read data from Amazon SQS.
Name | Type | Required | Default | Description |
---|---|---|---|---|
url | String | Yes | - | The Queue URL to read from Amazon SQS. |
region | String | No | - | The AWS region for the SQS service |
schema | Config | No | - | The structure of the data, including field names and field types. |
format | String | No | json | Data format. The default format is json. Optional text format, canal-json and debezium-json.If you use json or text format. The default field separator is ", ". If you customize the delimiter, add the “field_delimiter” option.If you use canal format, please refer to canal-json for details.If you use debezium format, please refer to debezium-json for details. |
format_error_handle_way | String | No | fail | The processing method of data format error. The default value is fail, and the optional value is (fail, skip). When fail is selected, data format error will block and an exception will be thrown. When skip is selected, data format error will skip this line data. |
field_delimiter | String | No | , | Customize the field delimiter for data format. |
common-options | No | - | Source plugin common parameters, please refer to Source Common Options for details |
source { AmazonSqs { url = "http://127.0.0.1:4566" region = "us-east-1" format = text field_delimiter = "#" schema = { fields { artist = string c_map = "map<string, array<int>>" c_array = "array<int>" c_string = string c_boolean = boolean c_tinyint = tinyint c_smallint = smallint c_int = int c_bigint = bigint c_float = float c_double = double c_decimal = "decimal(30, 8)" c_null = "null" c_bytes = bytes c_date = date c_timestamp = timestamp } } } } transform { # If you would like to get more information about how to configure seatunnel and see full list of transform plugins, # please go to https://seatunnel.apache.org/docs/transform-v2/sql } sink { Console {} }