On Monday, November 20, 2017 at 10:49:01 AM UTC-5, Jason wrote:
> a pipeline can be described as a sequence of functions that are applied to an
> input with each subsequent function getting the output of the preceding
> out = f6(f5(f4(f3(f2(f1(in))))))
> However this isn't very readable and does not support conditionals.
> Tensorflow has tensor-focused pipepines:
> fc1 = layers.fully_connected(x, 256, activation_fn=tf.nn.relu,
> fc2 = layers.fully_connected(fc1, 256, activation_fn=tf.nn.relu,
> out = layers.fully_connected(fc2, 10, activation_fn=None, scope='out')
> I have some code which allows me to mimic this, but with an implied parameter.
> def executePipeline(steps, collection_funcs = [map, filter, reduce]):
> results = None
> for step in steps:
> func = step
> params = step
> if func in collection_funcs:
> print func, params
> results = func(functools.partial(params,
> *params[1:]), results)
> print func
> if results is None:
> results = func(*params)
> results = func(*(params+(results,)))
> return results
> executePipeline( [
> (read_rows, (in_file,)),
> (map, (lower_row, field)),
> (stash_rows, ('stashed_file', )),
> (map, (lemmatize_row, field)),
> (vectorize_rows, (field, min_count,)),
> (evaluate_rows, (weights, None)),
> (recombine_rows, ('stashed_file', )),
> (write_rows, (out_file,))
> Which gets me close, but I can't control where rows gets passed in. In the
> above code, it is always the last parameter.
> I feel like I'm reinventing a wheel here. I was wondering if there's already
> something that exists?
Why do I want this? Because I'm tired of writing code that is locked away in a
bespoke function. I'd have an army of functions all slightly different in
functionality. I require flexibility in defining pipelines, and I don't want a
custom pipeline to require any low-level coding. I just want to feed a sequence
of functions to a script and have it process it. A middle ground between the
shell | operator and bespoke python code. Sure, I could write many binaries
bound by shell, but there are some things done far easier in python because of
its extensive libraries and it can exist throughout the execution of the
pipeline whereas any temporary persistence has to be though environment
variables or files.
Well after examining your feedback, it looks like Grapevine has 99% of the
concepts that I wanted to invent, even if the | operator seems a bit clunky. I
personally prefer the affluent interface convention. But this should work.
Kamaelia could also work, but it seems a little bit more grandiose.
Thanks everyone who chimed in!