blob: 37f8f45d9ce07beaed835e639098eca4d6069580 [file] [log] [blame]
// THIS FILE IS AUTOMATICALLY GENERATED: DO NOT EDIT
= image:kamelets/aws-s3-streaming-upload-sink.svg[] AWS S3 Streaming upload Sink
*Provided by: "Apache Software Foundation"*
*Support Level for this Kamelet is: "Preview"*
Upload data to AWS S3 in streaming upload mode.
== Configuration Options
The following table summarizes the configuration options available for the `aws-s3-streaming-upload-sink` Kamelet:
[width="100%",cols="2,^2,3,^2,^2,^3",options="header"]
|===
| Property| Name| Description| Type| Default| Example
| *accessKey {empty}* *| Access Key| The access key obtained from AWS.| string| |
| *bucketNameOrArn {empty}* *| Bucket Name| The S3 Bucket name or ARN.| string| |
| *keyName {empty}* *| Key Name| Setting the key name for an element in the bucket through endpoint parameter. In Streaming Upload, with the default configuration, this will be the base for the progressive creation of files.| string| |
| *region {empty}* *| AWS Region| The AWS region to connect to.| string| | `"eu-west-1"`
| *secretKey {empty}* *| Secret Key| The secret key obtained from AWS.| string| |
| autoCreateBucket| Autocreate Bucket| Setting the autocreation of the S3 bucket bucketName.| boolean| `false`|
| batchMessageNumber| Batch Message Number| The number of messages composing a batch in streaming upload mode| int| `10`|
| batchSize| Batch Size| The batch size (in bytes) in streaming upload mode| int| `"1000000"`|
| namingStrategy| Naming Strategy| The naming strategy to use in streaming upload mode. There are 2 enums and the value can be one of progressive, random| string| `"progressive"`|
| restartingPolicy| Restarting Policy| The restarting policy to use in streaming upload mode. There are 2 enums and the value can be one of override, lastPart| string| `"lastPart"`|
| streamingUploadMode| Streaming Upload Mode| Setting the Streaming Upload Mode| boolean| `true`|
| streamingUploadTimeout| Streaming Upload Timeout| While streaming upload mode is true, this option set the timeout to complete upload| long| |
|===
NOTE: Fields marked with ({empty}*) are mandatory.
== Usage
This section summarizes how the `aws-s3-streaming-upload-sink` can be used in various contexts.
=== Knative Sink
The `aws-s3-streaming-upload-sink` Kamelet can be used as Knative sink by binding it to a Knative object.
.aws-s3-streaming-upload-sink-binding.yaml
[source,yaml]
----
apiVersion: camel.apache.org/v1alpha1
kind: KameletBinding
metadata:
name: aws-s3-streaming-upload-sink-binding
spec:
source:
ref:
kind: InMemoryChannel
apiVersion: messaging.knative.dev/v1
name: mychannel
sink:
ref:
kind: Kamelet
apiVersion: camel.apache.org/v1alpha1
name: aws-s3-streaming-upload-sink
properties:
accessKey: "The Access Key"
bucketNameOrArn: "The Bucket Name"
keyName: "The Key Name"
region: "eu-west-1"
secretKey: "The Secret Key"
----
Make sure you have xref:latest@camel-k::installation/installation.adoc[Camel K installed] into the Kubernetes cluster you're connected to.
Save the `aws-s3-streaming-upload-sink-binding.yaml` file into your hard drive, then configure it according to your needs.
You can run the sink using the following command:
[source,shell]
----
kubectl apply -f aws-s3-streaming-upload-sink-binding.yaml
----
==== *Dependencies*
The Kamelet needs the following dependencies:
- camel:aws2-s3
- camel:kamelet
==== *Binding to Knative using the Kamel CLI:*
The procedure described above can be simplified into a single execution of the `kamel bind` command:
[source,shell]
----
kamel bind channel:mychannel aws-s3-streaming-upload-sink -p "sink.accessKey=The Access Key" -p "sink.bucketNameOrArn=The Bucket Name" -p "sink.keyName=The Key Name" -p "sink.region=eu-west-1" -p "sink.secretKey=The Secret Key"
----
This will create the KameletBinding under the hood and apply it to the current namespace in the cluster.
=== Kafka Sink
The `aws-s3-streaming-upload-sink` Kamelet can be used as Kafka sink by binding it to a Kafka topic.
.aws-s3-streaming-upload-sink-binding.yaml
[source,yaml]
----
apiVersion: camel.apache.org/v1alpha1
kind: KameletBinding
metadata:
name: aws-s3-streaming-upload-sink-binding
spec:
source:
ref:
kind: KafkaTopic
apiVersion: kafka.strimzi.io/v1beta1
name: my-topic
sink:
ref:
kind: Kamelet
apiVersion: camel.apache.org/v1alpha1
name: aws-s3-streaming-upload-sink
properties:
accessKey: "The Access Key"
bucketNameOrArn: "The Bucket Name"
keyName: "The Key Name"
region: "eu-west-1"
secretKey: "The Secret Key"
----
Ensure that you've installed https://strimzi.io/[Strimzi] and created a topic named `my-topic` in the current namespace.
Make also sure you have xref:latest@camel-k::installation/installation.adoc[Camel K installed] into the Kubernetes cluster you're connected to.
Save the `aws-s3-streaming-upload-sink-binding.yaml` file into your hard drive, then configure it according to your needs.
You can run the sink using the following command:
[source,shell]
----
kubectl apply -f aws-s3-streaming-upload-sink-binding.yaml
----
==== *Binding to Kafka using the Kamel CLI:*
The procedure described above can be simplified into a single execution of the `kamel bind` command:
[source,shell]
----
kamel bind kafka.strimzi.io/v1beta1:KafkaTopic:my-topic aws-s3-streaming-upload-sink -p "sink.accessKey=The Access Key" -p "sink.bucketNameOrArn=The Bucket Name" -p "sink.keyName=The Key Name" -p "sink.region=eu-west-1" -p "sink.secretKey=The Secret Key"
----
This will create the KameletBinding under the hood and apply it to the current namespace in the cluster.
==== Kamelet source file
Have a look at the following link:
https://github.com/apache/camel-kamelets/blob/main/aws-s3-streaming-upload-sink.kamelet.yaml
// THIS FILE IS AUTOMATICALLY GENERATED: DO NOT EDIT