Re: [VOTE] Accept CarbonData into the Apache Incubator

2016-05-26 Thread Jean-Baptiste Onofré

Hi Jim,

good point. Let me try to explain this "gap" regarding my discussion 
with the team:


1. Some people have been involved mostly in architecture and design more 
directly in code. That's why they are part of the initial committer 
list, whereas they didn't really provide "visible" code on github.


2. Some people are no more involved in the project. That's why they 
don't appear on the initial committer list.


Regards
JB

On 05/26/2016 05:45 PM, Jim Jagielski wrote:

I am trying to align the list of initial committers with
the list of current/active contributors, according to
Github, and I am seeing people proposed who have not
contributed anything and people NOT proposed who seem
to be kinda active...

Sooo. -0


On May 25, 2016, at 4:24 PM, Jean-Baptiste Onofré  wrote:

Hi all,

following the discussion thread, I'm now calling a vote to accept CarbonData 
into the Incubator.

​[ ] +1 Accept CarbonData into the Apache Incubator
[ ] +0 Abstain
[ ] -1 Do not accept CarbonData into the Apache Incubator, because ...

This vote is open for 72 hours.

The proposal follows, you can also access the wiki page:
https://wiki.apache.org/incubator/CarbonDataProposal

Thanks !
Regards
JB

= Apache CarbonData =

== Abstract ==

Apache CarbonData is a new Apache Hadoop native file format for faster 
interactive
query using advanced columnar storage, index, compression and encoding 
techniques
to improve computing efficiency, in turn it will help speedup queries an order 
of
magnitude faster over PetaBytes of data.

CarbonData github address: https://github.com/HuaweiBigData/carbondata

== Background ==

Huawei is an ICT solution provider, we are committed to enhancing customer 
experiences for telecom carriers, enterprises, and consumers on big data, In 
order to satisfy the following customer requirements, we created a new Hadoop 
native file format:

* Support interactive OLAP-style query over big data in seconds.
* Support fast query on individual record which require touching all fields.
* Fast data loading speed and support incremental load in period of minutes.
* Support HDFS so that customer can leverage existing Hadoop cluster.
* Support time based data retention.

Based on these requirements, we investigated existing file formats in the 
Hadoop eco-system, but we could not find a suitable solution that satisfying 
requirements all at the same time, so we start designing CarbonData.

== Rationale ==

CarbonData contains multiple modules, which are classified into two categories:

1. CarbonData File Format: which contains core implementation for file format 
such as columnar,index,dictionary,encoding+compression,API for reading/writing 
etc.
2. CarbonData integration with big data processing framework such as Apache 
Spark, Apache Hive etc. Apache Beam is also planned to abstract the execution 
runtime.

=== CarbonData File Format ===

CarbonData file format is a columnar store in HDFS, it has many features that a 
modern columnar format has, such as splittable, compression schema ,complex 
data type etc. And CarbonData has following unique features:

 Indexing 

In order to support fast interactive query, CarbonData leverage indexing 
technology to reduce I/O scans. CarbonData files stores data along with index, 
the index is not stored separately but the CarbonData file itself contains the 
index. In current implementation, CarbonData supports 3 types of indexing:

1. Multi-dimensional Key (B+ Tree index)
The Data block are written in sequence to the disk and within each data blocks 
each column block is written in sequence. Finally, the metadata block for the 
file is written with information about byte positions of each block in the 
file, Min-Max statistics index and the start and end MDK of each data block. 
Since, the entire data in the file is in sorted order, the start and end MDK of 
each data block can be used to construct a B+Tree and the file can be logically 
 represented as a B+Tree with the data blocks as leaf nodes (on disk) and the 
remaining non-leaf nodes in memory.
2. Inverted index
Inverted index is widely used in search engine. By using this index, it helps 
processing/query engine to do filtering inside one HDFS block. Furthermore, 
query acceleration for count distinct like operation is made possible when 
combining bitmap and inverted index in query time.
3. MinMax index
For all columns, minmax index is created so that processing/query engine can 
skip scan that is not required.

 Global Dictionary 

Besides I/O reduction, CarbonData accelerates computation by using global 
dictionary, which enables processing/query engines to perform all processing on 
encoded data without having to convert the data (Late Materialization). We have 
observed dramatic performance improvement for OLAP analytic scenario where 
table contains many columns in string data type. The data is converted back to 
the user readable form just before processing/query 

Re: [DISCUSS] Accept SensSoft into the Incubator

2016-05-26 Thread Alexander Bezzubov
The proposal looks very interesting,
wiki link is https://wiki.apache.org/incubator/SensSoftProposal

It's great to see community building efforts around open source usability
tools!

--
Alex


On Fri, May 27, 2016 at 8:11 AM, Mattmann, Chris A (3980) <
chris.a.mattm...@jpl.nasa.gov> wrote:

> Here, here.
>
> The team is well poised for Incubation and trying to grow hopefully
> a larger community here at the ASF.
>
> ++
> Chris Mattmann, Ph.D.
> Chief Architect
> Instrument Software and Science Data Systems Section (398)
> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
> Office: 168-519, Mailstop: 168-527
> Email: chris.a.mattm...@nasa.gov
> WWW:  http://sunset.usc.edu/~mattmann/
> ++
> Director, Information Retrieval and Data Science Group (IRDS)
> Adjunct Associate Professor, Computer Science Department
> University of Southern California, Los Angeles, CA 90089 USA
> WWW: http://irds.usc.edu/
> ++
>
>
>
>
>
>
>
>
>
>
> On 5/24/16, 12:19 PM, "Poore, Joshua C."  wrote:
>
> >Hi Ted!
> >
> >DARPA XDATA is ending next March. We currently have support on another
> DARPA contract that is good through next year, excluding potential options.
> We also have some commercial contracts that will leverage these open source
> projects. Some of this funding can be used for community support and open
> source product support, as well.
> >
> >SensSoft is also being backed heavily by Draper. Currently this means
> that our programs offices are aggressively pursuing new contracts that
> leverage this project. We are also working with internal leadership on
> internal research and development funding for SensSoft. This is how we are
> working to make sure that SensSoft survives and thrives beyond the XDATA
> program that spawned it.
> >
> >To summarize, Draper is a soft money operation (we’re a not-for-profit).
> However, we are pushing hard to support the community around SensSoft
> wherever possible and are considering options for how to fold in overhead
> earned on dollars brought in for SensSoft projects to support the SensSoft
> community. Draper believes that inclusion into the Apache Foundation will
> help increase our visibility, and help harden these projects in ways that
> will help generate more revenue to continually support and build upon the
> project.
> >
> >Thanks,
> >
> >Josh
> >
> >
> >Joshua C. Poore, Ph.D.
> >Senior Member of the Technical Staff
> >Draper
> >555 Technology Square, Rm. 2242
> >Cambridge, MA 02139-3563
> >Phone: (617) 258-4023
> >Cell: (617) 352-1700
> >Email: jpo...@draper.com
> >Participate in Operation XDATA: http://xdataonline.com!
> >
> >
> >
> >From: Ted Dunning [mailto:ted.dunn...@gmail.com]
> >Sent: Tuesday, May 24, 2016 9:44 AM
> >To: general@incubator.apache.org
> >Cc: Poore, Joshua C. 
> >Subject: Re: [DISCUSS] Accept SensSoft into the Incubator
> >
> >
> >This looks like an excellent project.
> >
> >How likely is it that it will be able to survive a hypothetical loss of
> DARPA funding?
> >
> >
> >
> >On Mon, May 23, 2016 at 3:00 PM, lewis john mcgibbney  > wrote:
> >Hi general@,
> >I would like to open a DISCUSS thread on the topic of accepting The
> >Software as a Sensor™ (SensSoft <
> https://wiki.apache.org/incubator/SensSoft>)
> >Project into the Incubator. I am CC'ing Joshua Poore from the Charles
> Stark
> >Draper Laboratory, Inc. who we have been working with to build community
> >around a kick-ass set of software projects under the SensSoft umbrella.
> >At this stage we would very much appreciate critical feedback from
> general@
> >community.
> >We are also open to mentors who may have an interest in the project
> >proposal.
> >The proposal is pasted below.
> >Thanks in advance,
> >Lewis
> >
> >= SensSoft Proposal =
> >
> >== Abstract ==
> >The Software as a Sensor™ (SensSoft) Project offers an open-source
> (ALv2.0)
> >software tool usability testing platform. It includes a number of
> >components that work together to provide a platform for collecting data
> >about user interactions with software tools, as well as archiving,
> >analyzing and visualizing that data. Additional components allow for
> >conducting web-based experiments in order to capture this data within a
> >larger experimental framework for formal user testing. These components
> >currently support Java Script-based web applications, although the schema
> >for “logging” user interactions can support mobile and desktop
> >applications, as well. Collectively, the Software as a Sensor Project
> >provides an open source platform for assessing how users interacted with
> >technology, not just collecting what they interacted with.
> >
> >== Proposal 

RE: [VOTE] Accept CarbonData into the Apache Incubator

2016-05-26 Thread Zheng, Kai
+1 (non-binding)

Regards,
Kai

-Original Message-
From: Gangumalla, Uma [mailto:uma.ganguma...@intel.com] 
Sent: Friday, May 27, 2016 1:10 AM
To: general@incubator.apache.org
Subject: Re: [VOTE] Accept CarbonData into the Apache Incubator

+1 (binding)

Regards,
Uma

On 5/25/16, 1:24 PM, "Jean-Baptiste Onofré"  wrote:

>Hi all,
>
>following the discussion thread, I'm now calling a vote to accept 
>CarbonData into the Incubator.
>
>​[ ] +1 Accept CarbonData into the Apache Incubator [ ] +0 Abstain [ ] 
>-1 Do not accept CarbonData into the Apache Incubator, because ...
>
>This vote is open for 72 hours.
>
>The proposal follows, you can also access the wiki page:
>https://wiki.apache.org/incubator/CarbonDataProposal
>
>Thanks !
>Regards
>JB
>
>= Apache CarbonData =
>
>== Abstract ==
>
>Apache CarbonData is a new Apache Hadoop native file format for faster 
>interactive query using advanced columnar storage, index, compression 
>and encoding techniques to improve computing efficiency, in turn it 
>will help speedup queries an order of magnitude faster over PetaBytes 
>of data.
>
>CarbonData github address: https://github.com/HuaweiBigData/carbondata
>
>== Background ==
>
>Huawei is an ICT solution provider, we are committed to enhancing 
>customer experiences for telecom carriers, enterprises, and consumers 
>on big data, In order to satisfy the following customer requirements, 
>we created a new Hadoop native file format:
>
>  * Support interactive OLAP-style query over big data in seconds.
>  * Support fast query on individual record which require touching all 
>fields.
>  * Fast data loading speed and support incremental load in period of 
>minutes.
>  * Support HDFS so that customer can leverage existing Hadoop cluster.
>  * Support time based data retention.
>
>Based on these requirements, we investigated existing file formats in 
>the Hadoop eco-system, but we could not find a suitable solution that 
>satisfying requirements all at the same time, so we start designing 
>CarbonData.
>
>== Rationale ==
>
>CarbonData contains multiple modules, which are classified into two
>categories:
>
>  1. CarbonData File Format: which contains core implementation for 
>file format such as columnar,index,dictionary,encoding+compression,API 
>for reading/writing etc.
>  2. CarbonData integration with big data processing framework such as 
>Apache Spark, Apache Hive etc. Apache Beam is also planned to abstract 
>the execution runtime.
>
>=== CarbonData File Format ===
>
>CarbonData file format is a columnar store in HDFS, it has many 
>features that a modern columnar format has, such as splittable, 
>compression schema ,complex data type etc. And CarbonData has following 
>unique
>features:
>
> Indexing 
>
>In order to support fast interactive query, CarbonData leverage 
>indexing technology to reduce I/O scans. CarbonData files stores data 
>along with index, the index is not stored separately but the CarbonData 
>file itself contains the index. In current implementation, CarbonData 
>supports 3 types of indexing:
>
>1. Multi-dimensional Key (B+ Tree index)
>  The Data block are written in sequence to the disk and within each 
>data blocks each column block is written in sequence. Finally, the 
>metadata block for the file is written with information about byte 
>positions of each block in the file, Min-Max statistics index and the 
>start and end MDK of each data block. Since, the entire data in the 
>file is in sorted order, the start and end MDK of each data block can 
>be used to construct a B+Tree and the file can be logically  
>represented as a
>B+Tree with the data blocks as leaf nodes (on disk) and the remaining
>non-leaf nodes in memory.
>2. Inverted index
>  Inverted index is widely used in search engine. By using this index, 
>it helps processing/query engine to do filtering inside one HDFS block.
>Furthermore, query acceleration for count distinct like operation is 
>made possible when combining bitmap and inverted index in query time.
>3. MinMax index
>  For all columns, minmax index is created so that processing/query 
>engine can skip scan that is not required.
>
> Global Dictionary 
>
>Besides I/O reduction, CarbonData accelerates computation by using 
>global dictionary, which enables processing/query engines to perform 
>all processing on encoded data without having to convert the data (Late 
>Materialization). We have observed dramatic performance improvement for 
>OLAP analytic scenario where table contains many columns in string data 
>type. The data is converted back to the user readable form just before 
>processing/query engine returning results to user.
>
> Column Group 
>
>Sometimes users want to perform processing/query on multi-columns in 
>one table, for example, performing scan for individual record in 
>troubleshooting scenario. In this case, row format is more efficient 
>than columnar format since all columns will be touched by the 

Re: [DISCUSS] Accept SensSoft into the Incubator

2016-05-26 Thread Mattmann, Chris A (3980)
Here, here.

The team is well poised for Incubation and trying to grow hopefully
a larger community here at the ASF.

++
Chris Mattmann, Ph.D.
Chief Architect
Instrument Software and Science Data Systems Section (398)
NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
Office: 168-519, Mailstop: 168-527
Email: chris.a.mattm...@nasa.gov
WWW:  http://sunset.usc.edu/~mattmann/
++
Director, Information Retrieval and Data Science Group (IRDS)
Adjunct Associate Professor, Computer Science Department
University of Southern California, Los Angeles, CA 90089 USA
WWW: http://irds.usc.edu/
++










On 5/24/16, 12:19 PM, "Poore, Joshua C."  wrote:

>Hi Ted!
>
>DARPA XDATA is ending next March. We currently have support on another DARPA 
>contract that is good through next year, excluding potential options. We also 
>have some commercial contracts that will leverage these open source projects. 
>Some of this funding can be used for community support and open source product 
>support, as well.
>
>SensSoft is also being backed heavily by Draper. Currently this means that our 
>programs offices are aggressively pursuing new contracts that leverage this 
>project. We are also working with internal leadership on internal research and 
>development funding for SensSoft. This is how we are working to make sure that 
>SensSoft survives and thrives beyond the XDATA program that spawned it.
>
>To summarize, Draper is a soft money operation (we’re a not-for-profit). 
>However, we are pushing hard to support the community around SensSoft wherever 
>possible and are considering options for how to fold in overhead earned on 
>dollars brought in for SensSoft projects to support the SensSoft community. 
>Draper believes that inclusion into the Apache Foundation will help increase 
>our visibility, and help harden these projects in ways that will help generate 
>more revenue to continually support and build upon the project.
>
>Thanks,
>
>Josh
>
>
>Joshua C. Poore, Ph.D.
>Senior Member of the Technical Staff
>Draper
>555 Technology Square, Rm. 2242
>Cambridge, MA 02139-3563
>Phone: (617) 258-4023
>Cell: (617) 352-1700
>Email: jpo...@draper.com
>Participate in Operation XDATA: http://xdataonline.com!
>
>
>
>From: Ted Dunning [mailto:ted.dunn...@gmail.com]
>Sent: Tuesday, May 24, 2016 9:44 AM
>To: general@incubator.apache.org
>Cc: Poore, Joshua C. 
>Subject: Re: [DISCUSS] Accept SensSoft into the Incubator
>
>
>This looks like an excellent project.
>
>How likely is it that it will be able to survive a hypothetical loss of DARPA 
>funding?
>
>
>
>On Mon, May 23, 2016 at 3:00 PM, lewis john mcgibbney 
>> wrote:
>Hi general@,
>I would like to open a DISCUSS thread on the topic of accepting The
>Software as a Sensor™ (SensSoft )
>Project into the Incubator. I am CC'ing Joshua Poore from the Charles Stark
>Draper Laboratory, Inc. who we have been working with to build community
>around a kick-ass set of software projects under the SensSoft umbrella.
>At this stage we would very much appreciate critical feedback from general@
>community.
>We are also open to mentors who may have an interest in the project
>proposal.
>The proposal is pasted below.
>Thanks in advance,
>Lewis
>
>= SensSoft Proposal =
>
>== Abstract ==
>The Software as a Sensor™ (SensSoft) Project offers an open-source (ALv2.0)
>software tool usability testing platform. It includes a number of
>components that work together to provide a platform for collecting data
>about user interactions with software tools, as well as archiving,
>analyzing and visualizing that data. Additional components allow for
>conducting web-based experiments in order to capture this data within a
>larger experimental framework for formal user testing. These components
>currently support Java Script-based web applications, although the schema
>for “logging” user interactions can support mobile and desktop
>applications, as well. Collectively, the Software as a Sensor Project
>provides an open source platform for assessing how users interacted with
>technology, not just collecting what they interacted with.
>
>== Proposal ==
>The Software as a Sensor™ Project is a next-generation platform for
>analyzing how individuals and groups of people make use of software tools
>to perform tasks or interact with other systems. It is composed of a number
>of integrated components:
> * User Analytic Logging Engine (User ALE) refers to a simple Application
>Program Interface (API) and backend infrastructure. User ALE provides
>“instrumentation” for software tools, such that each user interaction
>within the application can 

[VOTE] Apache SystemML 0.10.0-incubating (RC1)

2016-05-26 Thread Luciano Resende
Please vote to approve the release of the following candidate as Apache
SystemML version 0.10.0!

The PPMC vote thread:
https://www.mail-archive.com/dev@systemml.incubator.apache.org/msg00532.html

And the result:
https://www.mail-archive.com/dev@systemml.incubator.apache.org/msg00599.html

The tag to be voted on is v0.10.0-rc1
(ddf0e0941afe5d9c2cc7c574a6983aadd98c1fc3)

https://github.com/apache/incubator-systemml/tree/ddf0e0941afe5d9c2cc7c574a6983aadd98c1fc3

The release files, including signatures, digests, etc. can be found at:

https://repository.apache.org/content/repositories/orgapachesystemml-1005

The distribution and rat report is also available at:

https://dist.apache.org/repos/dist/dev/incubator/systemml/0.10.0-incubating-rc1/

The vote is open for at least 72 hours and passes if a majority of at least
3 +1 PMC votes are cast.

[ ] +1 Release this package as Apache SystemML 0.10.0
[ ] -1 Do not release this package because ...

-- 
Luciano Resende
http://twitter.com/lresende1975
http://lresende.blogspot.com/


[RESULT] [VOTE] Accept PredictionIO into the Apache Incubator

2016-05-26 Thread Andrew Purtell
​The ​
VOTE
​ to

​a
ccept PredictionIO into the Apache Incubator
​ has concluded and passed with 20 binding +1s, 8 non-binding +1s, and no 0
or -1 votes.

​Thanks to all who voted.

Binding +1s

Andrew Purtell
​Luciano Resende
James Taylor
Suneel Marthi
​Chris Nauroth
Roman Shaposhnik
Ted Dunning
Henry Saputra
​Drew Farris​
Jean-Baptiste Onofré
Uma Gangumalla
Sergio Fernández
John D. Ament
Ralph Goers
Seetharam Venkatesh
Hitesh Shah
Jake Farrell
Reynold Xin
Paul Fremantle
Bertrand Delacretaz

Nonbinding +1s

Ashish
Debo Dutta
Felix Cheung
Priyank Ashok Rastogi
Moon Soo Lee
Alexander Bezzubov
Tsuyoshi Ozawa
Sandeep Deshmukh

​No binding or nonbinding 0s

No binding or nonbinding -1s
​

--

Best regards,

   - Andy

Problems worthy of attack prove their worth by hitting back. - Piet Hein
(via Tom White)


Re: [VOTE] Accept CarbonData into the Apache Incubator

2016-05-26 Thread Gangumalla, Uma
+1 (binding)

Regards,
Uma

On 5/25/16, 1:24 PM, "Jean-Baptiste Onofré"  wrote:

>Hi all,
>
>following the discussion thread, I'm now calling a vote to accept
>CarbonData into the Incubator.
>
>​[ ] +1 Accept CarbonData into the Apache Incubator
>[ ] +0 Abstain
>[ ] -1 Do not accept CarbonData into the Apache Incubator, because ...
>
>This vote is open for 72 hours.
>
>The proposal follows, you can also access the wiki page:
>https://wiki.apache.org/incubator/CarbonDataProposal
>
>Thanks !
>Regards
>JB
>
>= Apache CarbonData =
>
>== Abstract ==
>
>Apache CarbonData is a new Apache Hadoop native file format for faster
>interactive
>query using advanced columnar storage, index, compression and encoding
>techniques
>to improve computing efficiency, in turn it will help speedup queries an
>order of
>magnitude faster over PetaBytes of data.
>
>CarbonData github address: https://github.com/HuaweiBigData/carbondata
>
>== Background ==
>
>Huawei is an ICT solution provider, we are committed to enhancing
>customer experiences for telecom carriers, enterprises, and consumers on
>big data, In order to satisfy the following customer requirements, we
>created a new Hadoop native file format:
>
>  * Support interactive OLAP-style query over big data in seconds.
>  * Support fast query on individual record which require touching all
>fields.
>  * Fast data loading speed and support incremental load in period of
>minutes.
>  * Support HDFS so that customer can leverage existing Hadoop cluster.
>  * Support time based data retention.
>
>Based on these requirements, we investigated existing file formats in
>the Hadoop eco-system, but we could not find a suitable solution that
>satisfying requirements all at the same time, so we start designing
>CarbonData.
>
>== Rationale ==
>
>CarbonData contains multiple modules, which are classified into two
>categories:
>
>  1. CarbonData File Format: which contains core implementation for file
>format such as columnar,index,dictionary,encoding+compression,API for
>reading/writing etc.
>  2. CarbonData integration with big data processing framework such as
>Apache Spark, Apache Hive etc. Apache Beam is also planned to abstract
>the execution runtime.
>
>=== CarbonData File Format ===
>
>CarbonData file format is a columnar store in HDFS, it has many features
>that a modern columnar format has, such as splittable, compression
>schema ,complex data type etc. And CarbonData has following unique
>features:
>
> Indexing 
>
>In order to support fast interactive query, CarbonData leverage indexing
>technology to reduce I/O scans. CarbonData files stores data along with
>index, the index is not stored separately but the CarbonData file itself
>contains the index. In current implementation, CarbonData supports 3
>types of indexing:
>
>1. Multi-dimensional Key (B+ Tree index)
>  The Data block are written in sequence to the disk and within each
>data blocks each column block is written in sequence. Finally, the
>metadata block for the file is written with information about byte
>positions of each block in the file, Min-Max statistics index and the
>start and end MDK of each data block. Since, the entire data in the file
>is in sorted order, the start and end MDK of each data block can be used
>to construct a B+Tree and the file can be logically  represented as a
>B+Tree with the data blocks as leaf nodes (on disk) and the remaining
>non-leaf nodes in memory.
>2. Inverted index
>  Inverted index is widely used in search engine. By using this index,
>it helps processing/query engine to do filtering inside one HDFS block.
>Furthermore, query acceleration for count distinct like operation is
>made possible when combining bitmap and inverted index in query time.
>3. MinMax index
>  For all columns, minmax index is created so that processing/query
>engine can skip scan that is not required.
>
> Global Dictionary 
>
>Besides I/O reduction, CarbonData accelerates computation by using
>global dictionary, which enables processing/query engines to perform all
>processing on encoded data without having to convert the data (Late
>Materialization). We have observed dramatic performance improvement for
>OLAP analytic scenario where table contains many columns in string data
>type. The data is converted back to the user readable form just before
>processing/query engine returning results to user.
>
> Column Group 
>
>Sometimes users want to perform processing/query on multi-columns in one
>table, for example, performing scan for individual record in
>troubleshooting scenario. In this case, row format is more efficient
>than columnar format since all columns will be touched by the workload.
>To accelerate this, CarbonData supports storing a group of column in row
>format, so data in column group is stored together and enable fast
>retrieval.
>
> Optimized for multiple use cases 
>
>CarbonData indices and dictionary is highly configurable. To make
>storage 

Re: [DISCUSS] CarbonData incubation proposal

2016-05-26 Thread Gangumalla, Uma
+1 (binding)


Regards,
Uma

On 5/18/16, 8:52 PM, "Jean-Baptiste Onofré"  wrote:

>Hi all,
>
>We would like to discuss about a new proposal for the incubator:
>CarbonData.
>
>CarbonData is a new Apache Hadoop native file format for faster
>interactive query using advanced columnar storage, index, compression
>and encoding techniques to improve computing efficiency, in turn it will
>help speedup queries an order of magnitude faster over PetaBytes of data.
>
>The proposal is included below and also available on the wiki:
>
>https://wiki.apache.org/incubator/CarbonDataProposal
>
>Please, provide any feedback or comment.
>
>Thanks !
>Regards
>JB
>
>= Apache CarbonData =
>
>== Abstract ==
>
>Apache CarbonData is a new Apache Hadoop native file format for faster
>interactive
>query using advanced columnar storage, index, compression and encoding
>techniques
>to improve computing efficiency, in turn it will help speedup queries an
>order of
>magnitude faster over PetaBytes of data.
>
>CarbonData github address: https://github.com/HuaweiBigData/carbondata
>
>== Backgrounad ==
>
>Huawei is an ICT solution provider, we are committed to enhancing
>customer experiences for telecom carriers, enterprises, and consumers on
>big data, In order to satisfy the following customer requirements, we
>created a new Hadoop native file format:
>
>  * Support interactive OLAP-style query over big data in seconds.
>  * Support fast query on individual record which require touching all
>fields.
>  * Fast data loading speed and support incremental load in period of
>minutes.
>  * Support HDFS so that customer can leverage existing Hadoop cluster.
>  * Support time based data retention.
>
>Based on these requirements, we investigated existing file formats in
>the Hadoop eco-system, but we could not find a suitable solution that
>satisfying requirements all at the same time, so we start designing
>CarbonData.
>
>== Rationale ==
>
>CarbonData contains multiple modules, which are classified into two
>categories:
>
>  1. CarbonData File Format: which contains core implementation for file
>format such as columnar,index,dictionary,encoding+compression,API for
>reading/writing etc.
>  2. CarbonData integration with big data processing framework such as
>Apache Spark, Apache Hive etc. Apache Beam is also planned to abstract
>the execution runtime.
>
>=== CarbonData File Format ===
>
>CarbonData file format is a columnar store in HDFS, it has many features
>that a modern columnar format has, such as splittable, compression
>schema ,complex data type etc. And CarbonData has following unique
>features:
>
> Indexing 
>
>In order to support fast interactive query, CarbonData leverage indexing
>technology to reduce I/O scans. CarbonData files stores data along with
>index, the index is not stored separately but the CarbonData file itself
>contains the index. In current implementation, CarbonData supports 3
>types of indexing:
>
>1. Multi-dimensional Key (B+ Tree index)
>  The Data block are written in sequence to the disk and within each
>data blocks each column block is written in sequence. Finally, the
>metadata block for the file is written with information about byte
>positions of each block in the file, Min-Max statistics index and the
>start and end MDK of each data block. Since, the entire data in the file
>is in sorted order, the start and end MDK of each data block can be used
>to construct a B+Tree and the file can be logically  represented as a
>B+Tree with the data blocks as leaf nodes (on disk) and the remaining
>non-leaf nodes in memory.
>2. Inverted index
>  Inverted index is widely used in search engine. By using this index,
>it helps processing/query engine to do filtering inside one HDFS block.
>Furthermore, query acceleration for count distinct like operation is
>made possible when combining bitmap and inverted index in query time.
>3. MinMax index
>  For all columns, minmax index is created so that processing/query
>engine can skip scan that is not required.
>
> Global Dictionary 
>
>Besides I/O reduction, CarbonData accelerates computation by using
>global dictionary, which enables processing/query engines to perform all
>processing on encoded data without having to convert the data (Late
>Materialization). We have observed dramatic performance improvement for
>OLAP analytic scenario where table contains many columns in string data
>type. The data is converted back to the user readable form just before
>processing/query engine returning results to user.
>
> Column Group 
>
>Sometimes users want to perform processing/query on multi-columns in one
>table, for example, performing scan for individual record in
>troubleshooting scenario. In this case, row format is more efficient
>than columnar format since all columns will be touched by the workload.
>To accelerate this, CarbonData supports storing a group of column in row
>format, so data in column group is stored together and enable fast

Re: [VOTE] Accept Pony Mail into the Apache Incubator

2016-05-26 Thread Jim Jagielski
+1 (binding)
> On May 24, 2016, at 1:56 AM, Daniel Gruno  wrote:
> 
> Since it seems the discussion has died down, I am now calling a vote on
> accepting Pony Mail into the Incubator. Sorry in advance for potato.
> 
> This vote will run for the usual 72 hours.
> 
> ### PROPOSAL BELOW ###
> 
> Abstract
> 
> Pony Mail is a mail-archiving, archive viewing, and interaction service,
> that can be integrated with many email platforms.
> 
> Proposal
> 
> Background
> 
> Pony Mail began as a response to two things; the lack of diversity in
> mailing list archives that are less bureaucratic all-or-nothing and more
> fluid way to interact with mailing lists than what is typically offered,
> and the lack of a performant system that solves this issue. Modern users
> of software want to jump right into a discussion they see, but cannot
> normally do so in a mailing list driven environment because of the rules
> generally surrounding said environment. Pony Mail, along with a select
> handful of newer archive systems, provides an interface that allows
> people to just hop into a thread, and take part. Without the need to
> subscribe, download the mbox archive, load it into your MTA, and respond.
> 
> As Rich writes in a very short essay:
> 
> You see a thread in which someone is WRONG ON THE INTERNET! You need to
> correct them. How do you do this today? You kinda don't. If you really
> wanted, you could download mbox files (and who the hell knows where they
> are?) and then try to get them into your mail client (which never works)
> and then reply to it. Which will break threading, because you did
> something wrong. Then you tear out your hair. PONY MAIL TO THE RESCUE!!!
> (sound of hoof beats)
> 
> Rationale
> 
> One of the oft-heard complaints about Apache's development model is that
> mailing lists are an old person's tool, and web-based communication -
> forums - are the way to go in the 21st Century. Providing a
> full-featured forum-like interface to mailing lists is one goal,while
> keeping all of the enormous benefits that mailing lists already provide.
> Asecond goal is to provide the ability to "jump in" to a mailing list
> conversation - even one that was a while back, without the convolutions
> that a mailing list requires. That is, to join this conversation the old
> way, one would have had to subscribe to the mailing list, download an
> mbox, and import it into ones mail client, in order that I be able to
> reply to this message with correct threading. With Pony Mail, one has to
> do none of those things, but can simply reply using the Web UI. To us,
> this is a HUGE benefit for building community. The requirement to jump
> through hoops to join a mailing list conversation drives away a lot of
> people (at least, anecdotally, it does) and if we can remove that
> barrier I think we'll have an easier time of drawing a new generation
> into our projects.
> 
> Initial Goals
> 
> The initial goals of transitioning to the ASF is to expand and grow both
> the Pony codebase and community, and ensure the project's continued
> growth and stability through forming a diverse and reliable community,
> in which the various facets of developers and contributors help keep the
> project up to date with latest developments and technical as well as
> social needs.
> 
> Current Status
> 
>Meritocracy:
> 
> The bulk of the code has been written by Daniel Gruno to date, but has
> had oversight from other committers, and mentors.
> 
>All members of the Pony project and wider community have a deep
> understanding and appreciation for the ASF meritocracy ideals, and are
> almost solely current ASF Members.
> 
>Community:
>The community is currently heavily focused within the ASF, and
> more specifically the Infrastructure group. This is to be expected given
> the nature of how the code came into existence in the first place. It
> should be noted that we have started reaching out to other groups who we
> know are using mailing list systems and therefore also rely on mailing
> list archive interfaces.
> 
>Core Developers:
> 
> Almost all core developers are ASF members, and are already intimately
> familiar with the Apache Way.
> 
>Alignment:
> 
> Pony will be very in line with ASF practices and processes as many of
> the founding members are long term ASF members and committers.
> 
> Known Risks
> 
>Orphaned products:
> 
> We are not aware of any issues with orphaned products related to this
> project.
> 
>Pony Mail relies on a set of CSS3 templates as well as some very stable
>programming languages. We have no reason to believe these would
> be orphaned or, should they become orphaned, that it would impact the
> development of the project.
> 
>Inexperience with Open Source:
>Most of the current committers are already ASF members and
> committers, we do not believe there to be any concerns around OSS
> inexperience.
> 
>   

Re: [VOTE] Accept CarbonData into the Apache Incubator

2016-05-26 Thread Jim Jagielski
I am trying to align the list of initial committers with
the list of current/active contributors, according to
Github, and I am seeing people proposed who have not
contributed anything and people NOT proposed who seem
to be kinda active...

Sooo. -0

> On May 25, 2016, at 4:24 PM, Jean-Baptiste Onofré  wrote:
> 
> Hi all,
> 
> following the discussion thread, I'm now calling a vote to accept CarbonData 
> into the Incubator.
> 
> ​[ ] +1 Accept CarbonData into the Apache Incubator
> [ ] +0 Abstain
> [ ] -1 Do not accept CarbonData into the Apache Incubator, because ...
> 
> This vote is open for 72 hours.
> 
> The proposal follows, you can also access the wiki page:
> https://wiki.apache.org/incubator/CarbonDataProposal
> 
> Thanks !
> Regards
> JB
> 
> = Apache CarbonData =
> 
> == Abstract ==
> 
> Apache CarbonData is a new Apache Hadoop native file format for faster 
> interactive
> query using advanced columnar storage, index, compression and encoding 
> techniques
> to improve computing efficiency, in turn it will help speedup queries an 
> order of
> magnitude faster over PetaBytes of data.
> 
> CarbonData github address: https://github.com/HuaweiBigData/carbondata
> 
> == Background ==
> 
> Huawei is an ICT solution provider, we are committed to enhancing customer 
> experiences for telecom carriers, enterprises, and consumers on big data, In 
> order to satisfy the following customer requirements, we created a new Hadoop 
> native file format:
> 
> * Support interactive OLAP-style query over big data in seconds.
> * Support fast query on individual record which require touching all fields.
> * Fast data loading speed and support incremental load in period of minutes.
> * Support HDFS so that customer can leverage existing Hadoop cluster.
> * Support time based data retention.
> 
> Based on these requirements, we investigated existing file formats in the 
> Hadoop eco-system, but we could not find a suitable solution that satisfying 
> requirements all at the same time, so we start designing CarbonData.
> 
> == Rationale ==
> 
> CarbonData contains multiple modules, which are classified into two 
> categories:
> 
> 1. CarbonData File Format: which contains core implementation for file format 
> such as columnar,index,dictionary,encoding+compression,API for 
> reading/writing etc.
> 2. CarbonData integration with big data processing framework such as Apache 
> Spark, Apache Hive etc. Apache Beam is also planned to abstract the execution 
> runtime.
> 
> === CarbonData File Format ===
> 
> CarbonData file format is a columnar store in HDFS, it has many features that 
> a modern columnar format has, such as splittable, compression schema ,complex 
> data type etc. And CarbonData has following unique features:
> 
>  Indexing 
> 
> In order to support fast interactive query, CarbonData leverage indexing 
> technology to reduce I/O scans. CarbonData files stores data along with 
> index, the index is not stored separately but the CarbonData file itself 
> contains the index. In current implementation, CarbonData supports 3 types of 
> indexing:
> 
> 1. Multi-dimensional Key (B+ Tree index)
> The Data block are written in sequence to the disk and within each data 
> blocks each column block is written in sequence. Finally, the metadata block 
> for the file is written with information about byte positions of each block 
> in the file, Min-Max statistics index and the start and end MDK of each data 
> block. Since, the entire data in the file is in sorted order, the start and 
> end MDK of each data block can be used to construct a B+Tree and the file can 
> be logically  represented as a B+Tree with the data blocks as leaf nodes (on 
> disk) and the remaining non-leaf nodes in memory.
> 2. Inverted index
> Inverted index is widely used in search engine. By using this index, it helps 
> processing/query engine to do filtering inside one HDFS block. Furthermore, 
> query acceleration for count distinct like operation is made possible when 
> combining bitmap and inverted index in query time.
> 3. MinMax index
> For all columns, minmax index is created so that processing/query engine can 
> skip scan that is not required.
> 
>  Global Dictionary 
> 
> Besides I/O reduction, CarbonData accelerates computation by using global 
> dictionary, which enables processing/query engines to perform all processing 
> on encoded data without having to convert the data (Late Materialization). We 
> have observed dramatic performance improvement for OLAP analytic scenario 
> where table contains many columns in string data type. The data is converted 
> back to the user readable form just before processing/query engine returning 
> results to user.
> 
>  Column Group 
> 
> Sometimes users want to perform processing/query on multi-columns in one 
> table, for example, performing scan for individual record in troubleshooting 
> scenario. In this case, row format is more 

Re: [VOTE] Accept CarbonData into the Apache Incubator

2016-05-26 Thread David E Jones

+1

-David (jonesde@a.o)


> On 25 May 2016, at 13:24, Jean-Baptiste Onofré  wrote:
> 
> Hi all,
> 
> following the discussion thread, I'm now calling a vote to accept CarbonData 
> into the Incubator.
> 
> ​[ ] +1 Accept CarbonData into the Apache Incubator
> [ ] +0 Abstain
> [ ] -1 Do not accept CarbonData into the Apache Incubator, because ...
> 
> This vote is open for 72 hours.
> 
> The proposal follows, you can also access the wiki page:
> https://wiki.apache.org/incubator/CarbonDataProposal
> 
> Thanks !
> Regards
> JB
> 
> = Apache CarbonData =
> 
> == Abstract ==
> 
> Apache CarbonData is a new Apache Hadoop native file format for faster 
> interactive
> query using advanced columnar storage, index, compression and encoding 
> techniques
> to improve computing efficiency, in turn it will help speedup queries an 
> order of
> magnitude faster over PetaBytes of data.
> 
> CarbonData github address: https://github.com/HuaweiBigData/carbondata
> 
> == Background ==
> 
> Huawei is an ICT solution provider, we are committed to enhancing customer 
> experiences for telecom carriers, enterprises, and consumers on big data, In 
> order to satisfy the following customer requirements, we created a new Hadoop 
> native file format:
> 
> * Support interactive OLAP-style query over big data in seconds.
> * Support fast query on individual record which require touching all fields.
> * Fast data loading speed and support incremental load in period of minutes.
> * Support HDFS so that customer can leverage existing Hadoop cluster.
> * Support time based data retention.
> 
> Based on these requirements, we investigated existing file formats in the 
> Hadoop eco-system, but we could not find a suitable solution that satisfying 
> requirements all at the same time, so we start designing CarbonData.
> 
> == Rationale ==
> 
> CarbonData contains multiple modules, which are classified into two 
> categories:
> 
> 1. CarbonData File Format: which contains core implementation for file format 
> such as columnar,index,dictionary,encoding+compression,API for 
> reading/writing etc.
> 2. CarbonData integration with big data processing framework such as Apache 
> Spark, Apache Hive etc. Apache Beam is also planned to abstract the execution 
> runtime.
> 
> === CarbonData File Format ===
> 
> CarbonData file format is a columnar store in HDFS, it has many features that 
> a modern columnar format has, such as splittable, compression schema ,complex 
> data type etc. And CarbonData has following unique features:
> 
>  Indexing 
> 
> In order to support fast interactive query, CarbonData leverage indexing 
> technology to reduce I/O scans. CarbonData files stores data along with 
> index, the index is not stored separately but the CarbonData file itself 
> contains the index. In current implementation, CarbonData supports 3 types of 
> indexing:
> 
> 1. Multi-dimensional Key (B+ Tree index)
> The Data block are written in sequence to the disk and within each data 
> blocks each column block is written in sequence. Finally, the metadata block 
> for the file is written with information about byte positions of each block 
> in the file, Min-Max statistics index and the start and end MDK of each data 
> block. Since, the entire data in the file is in sorted order, the start and 
> end MDK of each data block can be used to construct a B+Tree and the file can 
> be logically  represented as a B+Tree with the data blocks as leaf nodes (on 
> disk) and the remaining non-leaf nodes in memory.
> 2. Inverted index
> Inverted index is widely used in search engine. By using this index, it helps 
> processing/query engine to do filtering inside one HDFS block. Furthermore, 
> query acceleration for count distinct like operation is made possible when 
> combining bitmap and inverted index in query time.
> 3. MinMax index
> For all columns, minmax index is created so that processing/query engine can 
> skip scan that is not required.
> 
>  Global Dictionary 
> 
> Besides I/O reduction, CarbonData accelerates computation by using global 
> dictionary, which enables processing/query engines to perform all processing 
> on encoded data without having to convert the data (Late Materialization). We 
> have observed dramatic performance improvement for OLAP analytic scenario 
> where table contains many columns in string data type. The data is converted 
> back to the user readable form just before processing/query engine returning 
> results to user.
> 
>  Column Group 
> 
> Sometimes users want to perform processing/query on multi-columns in one 
> table, for example, performing scan for individual record in troubleshooting 
> scenario. In this case, row format is more efficient than columnar format 
> since all columns will be touched by the workload. To accelerate this, 
> CarbonData supports storing a group of column in row format, so data in 
> column group is stored together and enable fast 

Re: [VOTE] Accept CarbonData into the Apache Incubator

2016-05-26 Thread Jake Farrell
+1 (binding)

-Jake

On Wed, May 25, 2016 at 4:24 PM, Jean-Baptiste Onofré 
wrote:

> Hi all,
>
> following the discussion thread, I'm now calling a vote to accept
> CarbonData into the Incubator.
>
> ​[ ] +1 Accept CarbonData into the Apache Incubator
> [ ] +0 Abstain
> [ ] -1 Do not accept CarbonData into the Apache Incubator, because ...
>
> This vote is open for 72 hours.
>
> The proposal follows, you can also access the wiki page:
> https://wiki.apache.org/incubator/CarbonDataProposal
>
> Thanks !
> Regards
> JB
>
> = Apache CarbonData =
>
> == Abstract ==
>
> Apache CarbonData is a new Apache Hadoop native file format for faster
> interactive
> query using advanced columnar storage, index, compression and encoding
> techniques
> to improve computing efficiency, in turn it will help speedup queries an
> order of
> magnitude faster over PetaBytes of data.
>
> CarbonData github address: https://github.com/HuaweiBigData/carbondata
>
> == Background ==
>
> Huawei is an ICT solution provider, we are committed to enhancing customer
> experiences for telecom carriers, enterprises, and consumers on big data,
> In order to satisfy the following customer requirements, we created a new
> Hadoop native file format:
>
>  * Support interactive OLAP-style query over big data in seconds.
>  * Support fast query on individual record which require touching all
> fields.
>  * Fast data loading speed and support incremental load in period of
> minutes.
>  * Support HDFS so that customer can leverage existing Hadoop cluster.
>  * Support time based data retention.
>
> Based on these requirements, we investigated existing file formats in the
> Hadoop eco-system, but we could not find a suitable solution that
> satisfying requirements all at the same time, so we start designing
> CarbonData.
>
> == Rationale ==
>
> CarbonData contains multiple modules, which are classified into two
> categories:
>
>  1. CarbonData File Format: which contains core implementation for file
> format such as columnar,index,dictionary,encoding+compression,API for
> reading/writing etc.
>  2. CarbonData integration with big data processing framework such as
> Apache Spark, Apache Hive etc. Apache Beam is also planned to abstract the
> execution runtime.
>
> === CarbonData File Format ===
>
> CarbonData file format is a columnar store in HDFS, it has many features
> that a modern columnar format has, such as splittable, compression schema
> ,complex data type etc. And CarbonData has following unique features:
>
>  Indexing 
>
> In order to support fast interactive query, CarbonData leverage indexing
> technology to reduce I/O scans. CarbonData files stores data along with
> index, the index is not stored separately but the CarbonData file itself
> contains the index. In current implementation, CarbonData supports 3 types
> of indexing:
>
> 1. Multi-dimensional Key (B+ Tree index)
>  The Data block are written in sequence to the disk and within each data
> blocks each column block is written in sequence. Finally, the metadata
> block for the file is written with information about byte positions of each
> block in the file, Min-Max statistics index and the start and end MDK of
> each data block. Since, the entire data in the file is in sorted order, the
> start and end MDK of each data block can be used to construct a B+Tree and
> the file can be logically  represented as a B+Tree with the data blocks as
> leaf nodes (on disk) and the remaining non-leaf nodes in memory.
> 2. Inverted index
>  Inverted index is widely used in search engine. By using this index, it
> helps processing/query engine to do filtering inside one HDFS block.
> Furthermore, query acceleration for count distinct like operation is made
> possible when combining bitmap and inverted index in query time.
> 3. MinMax index
>  For all columns, minmax index is created so that processing/query engine
> can skip scan that is not required.
>
>  Global Dictionary 
>
> Besides I/O reduction, CarbonData accelerates computation by using global
> dictionary, which enables processing/query engines to perform all
> processing on encoded data without having to convert the data (Late
> Materialization). We have observed dramatic performance improvement for
> OLAP analytic scenario where table contains many columns in string data
> type. The data is converted back to the user readable form just before
> processing/query engine returning results to user.
>
>  Column Group 
>
> Sometimes users want to perform processing/query on multi-columns in one
> table, for example, performing scan for individual record in
> troubleshooting scenario. In this case, row format is more efficient than
> columnar format since all columns will be touched by the workload. To
> accelerate this, CarbonData supports storing a group of column in row
> format, so data in column group is stored together and enable fast
> retrieval.
>
>  Optimized for multiple use cases 
>

RE: [VOTE] Release Apache Fineract 0.1.2 (incubating)

2016-05-26 Thread Markus Geiß
Hey Mike,

hope this finds you well. ; o)

You're right, we are not depending on a GPL licensed library.
But LGPL (Hibernate) is treated the same way as GPL, so
we need to do some work on our side.

Best,

Markus

.::YAGNI likes a DRY KISS::.



> From: m...@vorburger.ch
> Date: Wed, 25 May 2016 18:57:05 +
> Subject: Re: [VOTE] Release Apache Fineract 0.1.2 (incubating)
> To: general@incubator.apache.org
>
> Hello,
>
> This thread appears to mix LGPL (Hibernate) and GPL (nothing used in
> Fineract AFAIK). They are not quite the same of course. It seems like even
> LGPL really is a problem then?
>
> Just thought it would be worth to be precise here.
>
> Regards,
> Michael
>
> On Sun, 22 May 2016, 00:56 Justin Mclean,  wrote:
>
>> Hi,
>>
>>> Sadly we still have an unwanted dependency bundled and we are using it
>> in the code.
>>> That's why we still have the GPL reference in the LICENSE file.
>>
>> I can’t see that you have any GPL code bundled in the source release, you
>> may depend on something that’s GPL license but that's another issue.
>>
>> Even that may be OK if it’s an optional dependancy, i.e. it only used to
>> run tests and/or most users of the software wouldn’t download it then
>> that’s OK. [1]
>>
>>> We were under the impression that we can 'release' incubation releases
>> with some
>>> issues and clean them out during the polling.
>>
>> That is correct but bundling/depending on GPL licensed software is a bit
>> more serious. One option is to ask on legal@dicuss and get permission to
>> make a one off release with a GPL dependancy.
>>
>>> [1] Given we use some Hibernate annotations we thought it is best to
>> bundle the license
>>> as long as we did not fixed that.
>>
>> You only need to put things in license if the code is bundled, I’m not
>> sure that annotating code would require a mention in license.
>>
>> Thanks,
>> Justin
>>
>> 1. http://www.apache.org/legal/resolved.html#optional
>> -
>> To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org
>> For additional commands, e-mail: general-h...@incubator.apache.org
>>
>>
  
-
To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org
For additional commands, e-mail: general-h...@incubator.apache.org



Re: Help with new ARIA TOSCA proposal

2016-05-26 Thread John D. Ament
Done, added you to the group.  Happy editing!

On Thu, May 26, 2016 at 7:40 AM Arthur Berezin 
wrote:

> My user ID is "ArthurBerezin" using my email art...@gigaspaces.com
>
> Thanks
>
> On Thu, May 26, 2016 at 12:12 AM John D. Ament 
> wrote:
>
>> Hi Arthur,
>>
>> If you send over your wiki username I can grant you access to start
>> putting up your proposal.
>>
>> John
>>
>> On Wed, May 25, 2016 at 4:57 PM Arthur Berezin 
>> wrote:
>>
>>> Sebastian, thanks for the kind introduction.
>>> John and Jakob thank you so much for stepping up! As Sebastian
>>> mentioned, we've been working for some time now preparing the code base we
>>> would like to be the starting point for ARIA as an Apache incubation
>>> project.
>>>
>>> I've tried to create a project proposal wiki page but it seems that I
>>> don't have needed permission to create new pages,
>>>
>>> I'd be happy to set a quick call sometime tomorrow or next week to
>>> discuss our motivation and current state, and also discuss the next
>>> steps we need to take in order to initiate the ASF incubation project.
>>>
>>> Best,
>>> Arthur
>>>
>>> On Wed, May 25, 2016 at 8:43 AM Jakob Homan  wrote:
>>>
 This is an area I'm focusing on right now.  I can sign up to mentor,
 if that would be useful.

 -Jakob


 On 24 May 2016 at 17:57, John D. Ament  wrote:
 > Hi,
 >
 > If you're still looking for help, I'd be happy to step up.
 >
 > John
 >
 > On Tue, May 24, 2016 at 8:34 AM Sebastien Goasguen 
 wrote:
 >
 >> Hi folks,
 >>
 >> I am an ASF member but I have not been involved with the incubator
 much.
 >>
 >> The last few weeks I have been chatting with folks at Gigaspaces
 [1]. Some
 >> of you may know them as the ones behind Cloudify [2] a cloud
 orchestration
 >> platform.
 >>
 >> The folks at Gigaspaces are interested to donate the orchestration
 engine
 >> behind Cloudify to ASF. With the motivation to open the governance
 of the
 >> core of Cloudify.
 >>
 >> They have already started laying a foundation for an incubator
 project,
 >> putting most of the software under a separate organization called
 ARIA [3]
 >> with software currently in gitHub [4].
 >>
 >> I would like to introduce Arthur Berezin from Gigaspaces who is
 working on
 >> putting together a proposal to the incubator. He got an account on
 the Wiki
 >> but no karma to add a page yet it seems.
 >>
 >> The proposal would need a champion, I don’t feel I have the
 experience
 >> with the incubator to do this well, so while I will help, it would
 be great
 >> to get a Champion to help shepherd this properly. Any takers ?
 >>
 >> Cheers,
 >>
 >> [1] http://www.gigaspaces.com
 >> [2] http://getcloudify.org
 >> [3] http://ariatosca.org
 >> [4] https://github.com/aria-tosca
 >>
 >> -Sebastien
 >> cloudstack, libcloud PMC
 >> -
 >> To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org
 >> For additional commands, e-mail: general-h...@incubator.apache.org
 >>
 >>

>>>


Re: Help with new ARIA TOSCA proposal

2016-05-26 Thread Arthur Berezin
My user ID is "ArthurBerezin" using my email art...@gigaspaces.com

Thanks

On Thu, May 26, 2016 at 12:12 AM John D. Ament 
wrote:

> Hi Arthur,
>
> If you send over your wiki username I can grant you access to start
> putting up your proposal.
>
> John
>
> On Wed, May 25, 2016 at 4:57 PM Arthur Berezin 
> wrote:
>
>> Sebastian, thanks for the kind introduction.
>> John and Jakob thank you so much for stepping up! As Sebastian
>> mentioned, we've been working for some time now preparing the code base we
>> would like to be the starting point for ARIA as an Apache incubation
>> project.
>>
>> I've tried to create a project proposal wiki page but it seems that I
>> don't have needed permission to create new pages,
>>
>> I'd be happy to set a quick call sometime tomorrow or next week to
>> discuss our motivation and current state, and also discuss the next
>> steps we need to take in order to initiate the ASF incubation project.
>>
>> Best,
>> Arthur
>>
>> On Wed, May 25, 2016 at 8:43 AM Jakob Homan  wrote:
>>
>>> This is an area I'm focusing on right now.  I can sign up to mentor,
>>> if that would be useful.
>>>
>>> -Jakob
>>>
>>>
>>> On 24 May 2016 at 17:57, John D. Ament  wrote:
>>> > Hi,
>>> >
>>> > If you're still looking for help, I'd be happy to step up.
>>> >
>>> > John
>>> >
>>> > On Tue, May 24, 2016 at 8:34 AM Sebastien Goasguen 
>>> wrote:
>>> >
>>> >> Hi folks,
>>> >>
>>> >> I am an ASF member but I have not been involved with the incubator
>>> much.
>>> >>
>>> >> The last few weeks I have been chatting with folks at Gigaspaces [1].
>>> Some
>>> >> of you may know them as the ones behind Cloudify [2] a cloud
>>> orchestration
>>> >> platform.
>>> >>
>>> >> The folks at Gigaspaces are interested to donate the orchestration
>>> engine
>>> >> behind Cloudify to ASF. With the motivation to open the governance of
>>> the
>>> >> core of Cloudify.
>>> >>
>>> >> They have already started laying a foundation for an incubator
>>> project,
>>> >> putting most of the software under a separate organization called
>>> ARIA [3]
>>> >> with software currently in gitHub [4].
>>> >>
>>> >> I would like to introduce Arthur Berezin from Gigaspaces who is
>>> working on
>>> >> putting together a proposal to the incubator. He got an account on
>>> the Wiki
>>> >> but no karma to add a page yet it seems.
>>> >>
>>> >> The proposal would need a champion, I don’t feel I have the experience
>>> >> with the incubator to do this well, so while I will help, it would be
>>> great
>>> >> to get a Champion to help shepherd this properly. Any takers ?
>>> >>
>>> >> Cheers,
>>> >>
>>> >> [1] http://www.gigaspaces.com
>>> >> [2] http://getcloudify.org
>>> >> [3] http://ariatosca.org
>>> >> [4] https://github.com/aria-tosca
>>> >>
>>> >> -Sebastien
>>> >> cloudstack, libcloud PMC
>>> >> -
>>> >> To unsubscribe, e-mail: general-unsubscr...@incubator.apache.org
>>> >> For additional commands, e-mail: general-h...@incubator.apache.org
>>> >>
>>> >>
>>>
>>


Accidentally sent podling report reminders for May and June

2016-05-26 Thread John D. Ament
Apologies to any podlings who may have been thrown off.  Please disregard
the reminders for May, as that was already completed.

JOhn