Here is the apache incubator guide on release - http://incubator.apache.org/guides/releasemanagement.html , http://wiki.apache.org/incubator/ReleaseChecklist As Ted mentioned, the first release would involve some setup and possibly few iterations to get it right. As you work on the release, you might want to create a singa how-to-release guide for making it easier to release in future. We have something like that for Hive - https://cwiki.apache.org/confluence/display/Hive/HowToRelease
On Thu, Aug 20, 2015 at 10:19 PM, Wang Wei <wang...@comp.nus.edu.sg> wrote: > Dear Mentors, > > We noticed a new proposal (Horn > https://wiki.apache.org/incubator/HornProposal) for developing a > distributed deep learning system. > Horn shares similar design ideas as SINGA. Beng Chin has given some details > below. > Maybe we can invite Horn developers to collaborate on SINGA? > If you know developers from Horn, you can introduce them to us. > Or if you have any other suggestions, please let us know. > > BTW, we are going to release the first version of SINGA. > Do you have any suggestions on the release? (Since this is the first > release, we are not quite clear on the process) > > Best, > Wei > > ---------- Forwarded message ---------- > From: ooibc <oo...@comp.nus.edu.sg> > Date: Fri, Aug 21, 2015 at 12:13 PM > Subject: Re: [DISCUSS] Horn Incubation Proposal > To: gene...@incubator.apache.org > Cc: "Edward J. Yoon" <edwardy...@apache.org> > > > > Hi, > > I am an initial committer of Apache(incubating) SINGA ( > http://singa.incubator.apache.org/) > > Both SINGA and the proposal follow the general parameter-server > architecture: > workers for computing gradients; servers for parameter updating. > > SINGA has implemented the model and data parallelism discussed in the Horn' > proposal: > multiple worker groups for asynchronous training---data parallelism; and > multiple workers in one group for synchronous training---model parallelism. > > One feature of SINGA's architecture is that it can be extended to organize > the > servers in a hierarchical topology, which may help to reduce the > communication bottleneck > of servers organized in a flat topology. > > For the programming model, currently Horn proposes to support feed-forward > models, > e.g., MLP, auto-encoder, while SINGA supports all three categories of the > known models, > feed-forward models (eg MLP, CNN), energy models (eg RBM, DBM), > and recurrent models (eg. RNN). > SINGA provides good support for users to code, e.g., implement new > parameter updating > protocols or layers, and is being integrated with HDFS as well. > > We will submit the first release and full documentation to the mentors this > weekend, and if > ok, we will announce the first full release soon. The GPU version is > scheduled for > October release. > > Technical papers: > http://www.comp.nus.edu.sg/~ooibc/singa-mm15.pdf > http://www.comp.nus.edu.sg/~ooibc/singaopen-mm15.pdf > > and project website (which has more details than the Apache web site): > http://www.comp.nus.edu.sg/~dbsystem/singa/ > > > There are plenty of rooms for collaborations indeed... > > regards > beng chin > www.comp.nus.edu.sg/~ooibc > > > > > On 2015-08-21 08:27, Edward J. Yoon wrote: > >> Hi all, >> >> We'd like to propose Horn (혼), a fully distributed system for >> large-scale deep learning as an Apache Incubator project and start the >> discussion. The complete proposal can be found at: >> https://wiki.apache.org/incubator/HornProposal >> >> Any advices and helps are welcome! Thanks, Edward. >> >> = Horn Proposal = >> >> == Abstract == >> >> (tentatively named "Horn [hɔ:n]", korean meaning of Horn is a >> "Spirit") is a neuron-centric programming APIs and execution framework >> for large-scale deep learning, built on top of Apache Hama. >> >> == Proposal == >> >> It is a goal of the Horn to provide a neuron-centric programming APIs >> which allows user to easily define the characteristic of artificial >> neural network model and its structure, and its execution framework >> that leverages the heterogeneous resources on Hama and Hadoop YARN >> cluster. >> >> == Background == >> >> The initial ANN code was developed at Apache Hama project by a >> committer, Yexi Jiang (Facebook) in 2013. The motivation behind this >> work is to build a framework that provides more intuitive programming >> APIs like Google's MapReduce or Pregel and supports applications >> needing large model with huge memory consumptions in distributed way. >> >> == Rationale == >> >> While many of deep learning open source softwares such as Caffe, >> DeepDist, and NeuralGiraph are still data or model parallel only, we >> aim to support both data and model parallelism and also fault-tolerant >> system design. The basic idea of data and model parallelism is use of >> the remote parameter server to parallelize model creation and >> distribute training across machines, and the BSP framework of Apache >> Hama for performing asynchronous mini-batches. Within single BSP job, >> each task group works asynchronously using region barrier >> synchronization instead of global barrier synchronization, and trains >> large-scale neural network model using assigned data sets in BSP >> paradigm. Thus, we achieve data and model parallelism. This >> architecture is inspired by Google's !DistBelief (Jeff Dean et al, >> 2012). >> >> == Initial Goals == >> >> Some current goals include: >> * builds new community >> * provides more intuitive programming APIs >> * needs both data and model parallelism support >> * must run natively on both Hama and Hadoop2 >> * needs also GPUs and InfiniBand support (FPGAs if possible) >> >> == Current Status == >> >> === Meritocracy === >> >> The core developers understand what it means to have a process based >> on meritocracy. We will provide continuous efforts to build an >> environment that supports this, encouraging community members to >> contribute. >> >> === Community === >> >> A small community has formed within the Apache Hama project and some >> companies such as instant messenger service company and mobile >> manufacturing company. And many people are interested in the >> large-scale deep learning platform itself. By bringing Horn into >> Apache, we believe that the community will grow even bigger. >> >> === Core Developers === >> >> Edward J. Yoon, Thomas Jungblut, and Dongjin Lee >> >> == Known Risks == >> >> === Orphaned Products === >> >> Apache Hama is already a core open source component at Samsung >> Electronics, and Horn also will be used by Samsung Electronics, and so >> there is no direct risk for this project to be orphaned. >> >> === Inexperience with Open Source === >> >> Some are very new and the others have experience using and/or working >> on Apache open source projects. >> >> === Homogeneous Developers === >> >> The initial committers are from different organizations such as, >> Microsoft, Samsung Electronics, and Line Plus. >> >> === Reliance on Salaried Developers === >> >> Few will be worked as a full-time open source developer. Other >> developers will also start working on the project in their spare time. >> >> === Relationships with Other Apache Products === >> >> * Horn is based on Apache Hama >> * Apache Zookeeper is used for distributed locking service >> * Natively run on Apache Hadoop and Mesos >> * Horn can be somewhat overlapped with Singa podling (If possible, >> we'd also like to use Singa or Caffe to do the heavy lifting part). >> >> === An Excessive Fascination with the Apache Brand === >> >> Horn itself will hopefully have benefits from Apache, in terms of >> attracting a community and establishing a solid group of developers, >> but also the relation with Apache Hama, a general-purpose BSP >> computing engine. These are the main reasons for us to send this >> proposal. >> >> == Documentation == >> >> Initial plan about Horn can be found at >> http://blog.udanax.org/2015/06/googles-distbelief-clone-project-on.html >> >> == Initial Source == >> >> The initial source code has been release as part of Apache Hama >> project developed under Apache Software Foundation. The source code is >> currently hosted at >> >> https://svn.apache.org/repos/asf/hama/trunk/ml/src/main/java/org/apache/hama/ml/ann/ >> >> == Cryptography == >> >> Not applicable. >> >> == Required Resources == >> >> === Mailing Lists === >> >> * horn-private >> * horn-dev >> >> === Subversion Directory === >> >> * Git is the preferred source control system: git://git.apache.org/horn >> >> === Issue Tracking === >> >> * a JIRA issue tracker, HORN >> >> == Initial Committers and Affiliations == >> >> * Thomas Jungblut (tjungblut AT apache DOT org) >> * Edward J. Yoon (edwardyoon AT apache DOT org) >> * Dongjin Lee (dongjin.lee.kr AT gmail DOT com) >> * Minho Kim (minwise.kim AT samsung DOT com) >> * Chia-Hung Lin (chl501 AT apache DOT org) >> * Behroz Sikander (behroz.sikander AT tum DOT de) >> * Hyok S. Choi (hyok.choi AT samsung DOT com) >> * Kisuk Lee (ks881115 AT gmail DOT com) >> >> == Affiliations == >> >> * Thomas Jungblut (Microsoft) >> * Edward J. Yoon (Samsung Electronics) >> * Donjin Lee (LINE Plus) >> * Minho Kim (Samsung Electronics) >> * Chia-Hung Lin (Self) >> * Behroz Sikander (Technical University of Munich) >> * Hyok S. Choi (Samsung Electronics) >> * Kisuk Lee (Seoul National University) >> >> == Sponsors == >> >> === Champion === >> >> * Edward J. Yoon <ASF member, edwardyoon AT apache DOT org> >> >> === Nominated Mentors === >> >> * Luciano Resende <ASF member, lresende AT apache DOT org> >> * Robin Anil <ASF member, robin.anil AT gmail DOT com> >> * Edward J. Yoon <ASF member, edwardyoon AT apache DOT org> >> >> === Sponsoring Entity === >> >> The Apache Incubator >> > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org > For additional commands, e-mail: general-h...@incubator.apache.org