Apache rocketmq

Clone this repo:
  1. 3acaed6 remove useless license header. by dinglei · 13 days ago main
  2. 1a309af Merge pull request #195 from ni-ze/Ibuilder by Ni Ze · 4 weeks ago
  3. ec7be1b reset to main by 维章 · 4 weeks ago
  4. edc4b5f remove dependency of channel-rocketmq in serviceLoader by 维章 · 4 weeks ago
  5. 2bc764d fix(ChannelBuilder) add CollectionSourceBuilder and add channel-db dependency into serviceloader by 维章 · 4 weeks ago

RocketMQ Streams

Build Status CodeCov GitHub release License Average time to resolve an issue Percentage of issues still open Twitter Follow


Quick Start


  • Lightweight deployment: RocketMQ Streams can be deployed separately or in cluster mode.
  • Various types of data input and output: source supports RocketMQ while sink supports databases and RocketMQ, etc.

Core API

RocketMQ Streams implements a series of advanced APIs, allowing users to write stream computing programs conveniently and achieve their own business requirements.


StreamBuilder is used to build the source of stream tasks.

DataStream API


DataStreamSource is a source class of segmented programming, used to interface with various data sources and obtain data from major message queues.

  • fromFile: reads data from the file. This method contains two parameters:

    • filePath: specifies which file path to read. Required.
    • isJsonData: specifies whether data is in JSON format. Optional. Default value: true.
    • tags: the tags for filtering messages used by the RocketMQ consumer. Optional.
  • fromRocketmq: obtains data from RocketMQ, including four parameters:

    • topic: the topic name of RocketMQ. Required.
    • groupName: the name of the consumer group. Required.
    • isJson: specifies whether data is in JSON format. Optional.
    • tags: the tags for filtering messages used by the RocketMQ consumer. Optional.
  • from: custom data source. You can specify your own data source by implementing ISource interface.


transform allows the input source data to be modified during the stream calculation process for the next step; DataStream API includes DataStream, JoinStream, SplitStream, WindowStream, and many other transform classes.


DataStream implements a series of common stream calculation operators as follows:

  • map: returns a new DataStream by passing each record of the source to the func function.
  • flatmap: similar to map. One input item corresponds to 0 or more output items.
  • filter: returns a new DataStream based on the record of the source DataStream only when the ** func** function returns true.
  • forEach: executes the func function once for each record and returns a new DataStream.
  • selectFields: returns the corresponding field value for each record, and returns a new DataStream.
  • operate: executes a custom function for each record and returns a new DataStream.
  • script: executes a script for each recorded field, returns new fields, and generates a new DataStream.
  • toPrint: prints the result on the console and generates a new DataStreamAction instance.
  • toFile: saves the result as a file and generates a new DataStreamAction instance.
  • toDB: saves the result to the database.
  • toRocketmq: outputs the result to RocketMQ.
  • to: outputs the result to the specified storage through the custom ISink interface.
  • window: performs relevant statistical analysis in the window, generally used in conjunction with groupBy. window() is used to define the size of the window, and groupBy( ) used to define the main key of statistical analysis. You can specify multiple main keys:
    • count: counts in the window.
    • min: gets the minimum of the statistical value in the window.
    • max: gets the maximum of the statistical value in the window.
    • avg: gets the average of the statistical values in the window.
    • sum: gets the sum of the statistical values in the window.
    • reduce: performs custom summary calculations in the window.
  • join: associates the two streams or one stream and one physical table according to the conditions and merges them into a large stream for related calculations.
    • dimJoin associate a stream with a physical table which can be a file or a db table, and all matching records are retained
    • dimLeftJoin After a flow is associated with a physical table, all data of the flow is reserved and fields that do not match the physical table are left blank
    • join
    • leftJoin
  • union: merges the two streams.
  • split: splits a data stream into different data streams according to tags for downstream analysis and calculation.
  • with: specifies related strategies during the calculation, including Checkpoint and state storage strategies, etc.


The Strategy mechanism is mainly used to control the underlying logic during the operation of the computing engine, such as the storage methods of window. Subsequent controls for state, dual-stream joins, and so on will be added. All control strategies are transmitted through the with operator. Multiple policy types can be transmitted at the same time.

//Specify the storage strategy for Checkpoint.

Run local project


  • JDK 1.8+
  • Maven 3.2+
  • Install RocketMQ in local,intall doc

Install Rocketmq-streams

git clone https://github.com/apache/rocketmq-streams.git
cd rocketmq-streams
mvn clean -DskipTests install -U


          <!--Newest version-->