Hi folks,

Falcon offers many services OOTB and caters to a wide array of use cases.
However, there has been many asks that does not fit the functionality
offered by Falcon. I'm proposing that we add recipes to Falcon which is
similar to recipes in Whirr and other management solutions such as puppet
and chef.

Overview:
A recipe essentially is a static process template with parameterized
workflow to realize a specific use case. For example:

* replicating directories from one HDFS cluster to another (not timed
partitions)
* replicating hive metadata (database, table, views, etc.)
* replicating between HDFS and Hive - either way
* anonymization of data based on schema
* data masking
* etc.

Proposal:
Falcon provides a Process abstraction that encapsulates the configuration
for a user workflow with scheduling controls. All recipes can be modeled
as a Process with in Falcon which executes the user workflow
periodically. The process and its associated workflow are parameterized.
The user will provide a properties file with name value pairs that are
substituted by falcon before scheduling it.

The CLI would look something like this:
falcon -recipe $recipe_name -properties $properties_file

Thoughts?

-- 
Regards,
Venkatesh

“Perfection (in design) is achieved not when there is nothing more to add,
but rather when there is nothing more to take away.”
- Antoine de Saint-Exupéry

Reply via email to