Hello Samza is a starter project for Apache Samza jobs.
Hello Samza is developed as part of the Apache Samza project. Please direct questions, improvements and bug fixes there. Questions about Hello Samza are welcome on the dev list and the Samza JIRA has a hello-samza component for filing tickets.
The Hello Samza project contains example Samza applications of high-level API as well as low-level API. The following are the instructions to install the binaries and run the applications in a local Yarn cluster. See also Hello Samza and Hello Samza High Level API for more information.
Check out the hello-samza project:
git clone https://gitbox.apache.org/repos/asf/samza-hello-samza.git hello-samza cd hello-samza
To build hello-samza with the latest Samza master, you can switch to the latest branch:
git checkout latest
This project contains everything you'll need to run your first Samza application.
This command will download, install, and start ZooKeeper, Kafka, and YARN. It will also check out the latest version of Samza and build it. All package files will be put in a sub-directory called “deploy” inside hello-samza's root folder.
If you get a complaint that JAVA_HOME is not set, then you'll need to set it to the path where Java is installed on your system.
Once the grid command completes, you can verify that YARN is up and running by going to http://localhost:8088. This is the YARN UI.
Before you can run a Samza application, you need to build a package for it. This package is what YARN uses to deploy your apps on the grid. Use the following command in hello-samza project to build and deploy the example applications:
After you've built your Samza package, you can start the example applications on the grid.
Package samza.examples.cookbook contains various examples of high-level API operator usage, such as map, partitionBy, window and join. Each example is a runnable Samza application with the steps in the class javadocs, e.g PageViewAdClickJoiner.
Package samza.examples.wikipedia.application contains a small Samza application which consumes the real-time feeds from Wikipedia, extracts the metadata of the events, and calculates statistics of all edits in a 10-second window. You can start the app on the grid using the run-app.sh script:
Once the job is started, we can tail the kafka topic by:
./deploy/kafka/bin/kafka-console-consumer.sh --bootstrap-server localhost:9092 --topic wikipedia-stats
A code walkthrough of this application can be found here.
Package samza.examples.wikipedia.task contains the low-level API Samza code for the Wikipedia example. To run it, use the following scripts:
deploy/samza/bin/run-app.sh --config-path=$PWD/deploy/samza/config/wikipedia-feed.properties deploy/samza/bin/run-app.sh --config-path=$PWD/deploy/samza/config/wikipedia-parser.properties deploy/samza/bin/run-app.sh --config-path=$PWD/deploy/samza/config/wikipedia-stats.properties
Once the jobs are started, you can use the same kafka-console-consumer.sh command as in the high-level API Wikipedia example to check out the output of the statistics.
Every example above are ran with a few messages as Integration test using TestRunner API. You can find all the testing samples in src/test/java. To run it use:
mvn clean package
Run Single example as test use:
mvn test -Dtest=<ClassName>