Hi Nayomi,

Current approach is an acceptable solution, where copying the DAS pack to a
target Spark node is not a problem. Having a script or something to compile
the list of jars only required for it is just a small convenience
mechanism, and is not really that important. Someone may actually just like
copying the pack right away, rather than running another command.

Cheers,
Anjana.

On Mon, Oct 19, 2015 at 10:24 PM, Nayomi Dayarathne <[email protected]> wrote:

> Hi all,
>
> Given the current situation in DAS, there should be a DAS distribution in
> every spark node in an external spark cluster in order to work DAS with
> it.That way, external spark cluster able to access jar files which needs to
> work with DAS.
>
> Since this is not a better approach,I have already reported an improvement
> JIRA[1] regarding this.
>
> Therefore, we want to know whether we are going to implement a solution
> according to the JIRA[1] reported or are there any better solution for this
> ?
>
>
> [1].https://wso2.org/jira/browse/DAS-197
>
>
> Regards,
>
> Nayomi Dayarathne
>
> *Software Engineer-QA*
> Mobile : +94 (0) 775246619 <+94+(0)+775246619>
>
> *[email protected] <[email protected]>*
>



-- 
*Anjana Fernando*
Senior Technical Lead
WSO2 Inc. | http://wso2.com
lean . enterprise . middleware
_______________________________________________
Dev mailing list
[email protected]
http://wso2.org/cgi-bin/mailman/listinfo/dev

Reply via email to