commit | 0825d80a638652bc68119425e910f969602c83c6 | [log] [tgz] |
---|---|---|
author | Pallavi Rao <pallavi.rao@inmobi.com> | Thu Feb 02 17:39:23 2017 +0530 |
committer | Pallavi Rao <pallavi.rao@inmobi.com> | Thu Feb 02 17:39:23 2017 +0530 |
tree | f08a2b9dc85c05eb18453591c3b9537e086375d1 | |
parent | 993861c850ba38d7a3d7313c290ec614c071928f [diff] |
FALCON-2267 Definition api fails if resources are empty This is extension of sandeepSamudrala's work on PR 353. Completing it on his behalf as he is out on vacation. Dev testing done: `IM1738M1:falcon-0.11-SNAPSHOT pallavi.rao$ bin/falcon extension -enumerate [ { "name": "sample", "type": "Custom extension", "location": "hdfs://192.168.138.236:8020/tmp/extensions/extension-example" }, ….. { "name": "hive-mirroring", "type": "Trusted extension", "description": "This extension implements replicating hive metadata and data from one Hadoop cluster to another Hadoop cluster.", "location": "file:/Users/pallavi.rao/falcon/falcon-0.11-SNAPSHOT/extensions/hive-mirroring" } ] IM1738M1:falcon-0.11-SNAPSHOT pallavi.rao$ bin/falcon extension -definition -extensionName hive-mirroring { "shortDescription":"This extension implements replicating hive metadata and data from one Hadoop cluster to another Hadoop cluster.", "properties":[ { "propertyName":"jobName", "required":true, "description":"Unique job name", "example":"hive-monthly-sales-dr" }, …. ] } IM1738M1:falcon-0.11-SNAPSHOT pallavi.rao$ bin/falcon extension -definition -extensionName sample Contents of file config: Contents of file config2: <workflow-app xmlns="uri:oozie:workflow:0.1" name="merlin-workflow"> …. IM1738M1:falcon-0.11-SNAPSHOT pallavi.rao$ bin/falcon extension -describe -extensionName sample Extension Test IM1738M1:falcon-0.11-SNAPSHOT pallavi.rao$ bin/falcon extension -describe -extensionName hive-mirroring ..... Hive Mirroring Extension Overview Falcon provides feature to replicate Hive metadata and data events from source cluster to destination cluster. …..` Author: Pallavi Rao <pallavi.rao@inmobi.com> Reviewers: @PracheerAgarwal, @sandeepSamudrala Closes #356 from pallavi-rao/2267
Falcon is a feed processing and feed management system aimed at making it easier for end consumers to onboard their feed processing and feed management on hadoop clusters.
Dependencies across various data processing pipelines are not easy to establish. Gaps here typically leads to either incorrect/partial processing or expensive reprocessing. Repeated duplicate definition of a single feed multiple times can lead to inconsistencies / issues.
Input data may not arrive always on time and it is required to kick off the processing without waiting for all data to arrive and accommodate late data separately
Feed management services such as feed retention, replications across clusters, archival etc are tasks that are burdensome on individual pipeline owners and better offered as a service for all customers.
It should be easy to onboard new workflows/pipelines
Smoother integration with metastore/catalog
Provide notification to end customer based on availability of feed groups (logical group of related feeds, which are likely to be used together)
You can find the documentation on Apache Falcon website.
Before opening a pull request, please go through the Contributing to Apache Falcon wiki. It lists steps that are required before creating a PR and the conventions that we follow. If you are looking for issues to pick up then you can look at starter tasks or open tasks
You can download release notes of previous releases from the following links.