hi folks, How would you like to proceed on this? I'm tracking many projects right now so I want to make sure someone else is "in charge" on this effort
Thanks, Wes On Sat, Oct 6, 2018 at 10:37 AM Wes McKinney <wesmck...@gmail.com> wrote: > > > We could create a worker pool like abstraction where the workers are the CI > > services, but that would require a scheduler to poll the finished jobs then > > submit the dependent ones. This sounds a bit inconvenient, where would that > > scheduler run: locally, on a CI or self hosted? > > Inevitably we're going to need to build some kind of job scheduler, > whether it uses Airflow or Luigi or some other tool of our own > devising. > > Apache Arrow is eventually going to need a host where we can manage > such workflows. I'm looking into the possibility of a physical > CUDA-equipped host that could be made available to Arrow developers to > use for testing and benchmarking. I may need to run the machine out of > my home (we did something similar for pandas -- physical machine that > we can SSH into). > > All this idealism aside -- we take the shortest path possible for this > particular packaging job, and make improvements as we can going > forward. > On Sat, Oct 6, 2018 at 9:31 AM Krisztián Szűcs > <szucs.kriszt...@gmail.com> wrote: > > > > I see now, so the jar would contain all of the three shared libraries. > > > > We could create a worker pool like abstraction where the workers are the > > CI services, but that would require a scheduler to poll the finished jobs > > then > > submit the dependent ones. This sounds a bit inconvenient, where would > > that scheduler run: locally, on a CI or self hosted? > > > > Another approach would be to use the worker the schedule the next task, > > in a similar fashion like dask's worker_client [1] launches tasks from > > tasks. > > There could be synchronization problems though. This approach requires > > to bootstrap crossbow on each CI jobs but that would: > > - make crossbow less CI dependent (to use azure pipelines as well) > > - unify the artifact uploading and downloading logic which is required in > > order > > to support dependent tasks > > - way less redundancy in task definitions > > > > What do You think? I'd prefer the second one. > > > > [1] > > https://github.com/dask/distributed/blob/master/docs/source/task-launch.rst > > > > On Sat, Oct 6, 2018 at 10:57 AM Wes McKinney <wesmck...@gmail.com> wrote: > > > > > It seems the complicated part of this will be having a dependent task > > > that packages up the 3 shared libraries, one for each platform, after > > > the individual packaging tasks are run. How would you propose handling > > > that? > > > On Fri, Oct 5, 2018 at 8:03 AM Krisztián Szűcs > > > <szucs.kriszt...@gmail.com> wrote: > > > > > > > > Ohh, just read the thread, sorry! > > > > > > > > So crossbow is located here > > > https://github.com/apache/arrow/tree/master/dev/tasks > > > > I suggest to "fork" the python-wheels directory which contains three > > > templated ymls > > > > for osx, win and linux builds. For building on linux something like the > > > following should > > > > be sufficient > > > https://gist.github.com/kszucs/39154876d60c4109ff59b678afd65b19 > > > > Then You need another entry in the tasks.yml, for example: > > > > jar-gandiva-linux: > > > > platform: linux > > > > template: gandiva-jars/travis.linux.yml > > > > params: > > > > # arbitrary params which are available from the templated yml > > > > ... > > > > artifacts: > > > > # these are the expected artifacts from the build > > > > - gandiva-SNAPSHOT-{version}.jar > > > > ... > > > > > > > > Of course crossbow is wired towards the current packaging requirements, > > > so likely > > > > We need to adjust it to the newly appearing requirements. > > > > > > > > Feel free to reach me on gitter @kszucs. > > > > On Oct 4 2018, at 2:02 pm, Wes McKinney <wesmck...@gmail.com> wrote: > > > > > > > > > > hi Praveen, > > > > > Probably the best way to accomplish this is to use our new Crossbow > > > > > infrastructure for task automation on Travis CI and Appveyor rather > > > > > than trying to do all of this within the CI entries. This is how we > > > > > are producing all of our binary artifacts for releases now -- > > > > > presumably in future ASF releases, we will want to include a > > > > > platform-independent Gandiva JAR in our release votes, so this all > > > > > needs to end up in Crossbow anyway. The intent is for the Crossbow > > > > > system to take on responsibility for all packaging automation rather > > > > > than using the normal CI for that. > > > > > > > > > > Krisztian, do you have time to help Praveen and the Gandiva crew with > > > > > this project? This will be an important test to document and improve > > > > > Crossbow for such use cases > > > > > > > > > > Thanks > > > > > Wes > > > > > On Thu, Oct 4, 2018 at 7:14 AM Praveen Kumar <prav...@dremio.com> > > > wrote: > > > > > > > > > > > > Hi Folks, > > > > > > As part of https://issues.apache.org/jira/browse/ARROW-3385, we are > > > > > > planning to perform a snapshot release of the Gandiva Jar on each > > > commit to > > > > > > master. This would be a platform independent jar that contains the > > > core > > > > > > gandiva library and its jni bridge packaged for Mac, Windows and > > > > > > *nix > > > > > > platforms. > > > > > > > > > > > > The current plan is to deploy separate snapshot jars for each OS > > > through > > > > > > entries in the Gandiva CI matrix and then have a combine step that > > > pulls in > > > > > > each OS specific jar and builds a jar that has all the native > > > libraries. > > > > > > This build/deploy would happen only for commits on master branch and > > > not > > > > > > for PR requests > > > > > > > > > > > > Does the plan sound ok (or) please let us know if there is a better > > > way to > > > > > > achieve the same. > > > > > > > > > > > > If it sounds ok, can someone please help with the following > > > > > > 1. It looks like we only do travis builds and not appveyor for > > > master in > > > > > > arrow. Any reason for this? > > > > > > 2. Even if we did appveyor is there a way to sequence the builds. > > > Like wait > > > > > > for appveyor to complete before kicking off travis? Since we would > > > need the > > > > > > dll to be pre-built. > > > > > > 3. Someone would need to configure the credentials to use for the > > > ossrh > > > > > > deployment. The credentials would need access to deploy to > > > org.apache.arrow. > > > > > > > > > > > > Thanks ahead! > > >