[jira] [Updated] (FLINK-8828) Add collect method to DataStream / DataSet scala api

2020-02-05 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-8828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-8828:
--
Labels: pull-request-available  (was: )

> Add collect method to DataStream / DataSet scala api
> 
>
> Key: FLINK-8828
> URL: https://issues.apache.org/jira/browse/FLINK-8828
> Project: Flink
>  Issue Type: Improvement
>  Components: API / DataSet, API / DataStream, API / Scala
>Affects Versions: 1.4.0
>Reporter: Jelmer Kuperus
>Priority: Major
>  Labels: pull-request-available
>
> A collect function is a method that takes a Partial Function as its parameter 
> and applies it to all the elements in the collection to create a new 
> collection which satisfies the Partial Function.
> It can be found on all [core scala collection 
> classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
>  as well as on spark's [rdd 
> interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]
> To understand its utility imagine the following scenario :
> Given a DataStream that produces events of type _Purchase_ and _View_ 
> Transform this stream into a stream of purchase amounts over 1000 euros.
> Currently an implementation might look like
> {noformat}
> val x = dataStream
>   .filter(_.isInstanceOf[Purchase])
>   .map(_.asInstanceOf[Purchase])
>   .filter(_.amount > 1000)
>   .map(_.amount){noformat}
> Or alternatively you could do this
> {noformat}
> dataStream.flatMap(_ match {
>   case p: Purchase if p.amount > 1000 => Some(p.amount)
>   case _ => None
> }){noformat}
> But with collect implemented it could look like
> {noformat}
> dataStream.collect {
>   case p: Purchase if p.amount > 1000 => p.amount
> }{noformat}
>  
> Which is a lot nicer to both read and write



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-8828) Add collect method to DataStream / DataSet scala api

2018-03-01 Thread Jelmer Kuperus (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-8828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jelmer Kuperus updated FLINK-8828:
--
Description: 
A collect function is a method that takes a Partial Function as its parameter 
and applies it to all the elements in the collection to create a new collection 
which satisfies the Partial Function.

It can be found on all [core scala collection 
classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
 as well as on spark's [rdd 
interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]

To understand its utility imagine the following scenario :

Given a DataStream that produces events of type _Purchase_ and _View_ 
Transform this stream into a stream of purchase amounts over 1000 euros.

Currently an implementation might look like
{noformat}
val x = dataStream
  .filter(_.isInstanceOf[Purchase])
  .map(_.asInstanceOf[Purchase])
  .filter(_.amount > 1000)
  .map(_.amount){noformat}
Or alternatively you could do this
{noformat}
dataStream.flatMap(_ match {
  case p: Purchase if p.amount > 1000 => Some(p.amount)
  case _ => None
}){noformat}
But with collect implemented it could look like
{noformat}
dataStream.collect {
  case p: Purchase if p.amount > 1000 => p.amount
}{noformat}
 

Which is a lot nicer to both read and write

  was:
A collect function is a method that takes a Partial Function as its parameter 
and applies it to all the elements in the collection to create a new collection 
which satisfies the Partial Function.

It can be found on all [core scala collection 
classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
 as well as on spark's [rdd 
interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]

To understand its utility imagine the following scenario :

You have a DataStream that produces events of type _Purchase_ and _View_ 
 You would like to transform this stream into a stream of purchase amounts over 
1000 euros.

Currently an implementation might look like
{noformat}
val x = dataStream
  .filter(_.isInstanceOf[Purchase])
  .map(_.asInstanceOf[Purchase])
  .filter(_.amount > 1000)
  .map(_.amount){noformat}
Or alternatively you could do this
{noformat}
dataStream.flatMap(_ match {
  case p: Purchase if p.amount > 1000 => Some(p.amount)
  case _ => None
}){noformat}
But with collect implemented it could look like
{noformat}
dataStream.collect {
  case p: Purchase if p.amount > 1000 => p.amount
}{noformat}
 

Which is a lot nicer to both read and write


> Add collect method to DataStream / DataSet scala api
> 
>
> Key: FLINK-8828
> URL: https://issues.apache.org/jira/browse/FLINK-8828
> Project: Flink
>  Issue Type: Improvement
>  Components: Core, DataSet API, DataStream API, Scala API
>Affects Versions: 1.4.0
>Reporter: Jelmer Kuperus
>Priority: Major
>
> A collect function is a method that takes a Partial Function as its parameter 
> and applies it to all the elements in the collection to create a new 
> collection which satisfies the Partial Function.
> It can be found on all [core scala collection 
> classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
>  as well as on spark's [rdd 
> interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]
> To understand its utility imagine the following scenario :
> Given a DataStream that produces events of type _Purchase_ and _View_ 
> Transform this stream into a stream of purchase amounts over 1000 euros.
> Currently an implementation might look like
> {noformat}
> val x = dataStream
>   .filter(_.isInstanceOf[Purchase])
>   .map(_.asInstanceOf[Purchase])
>   .filter(_.amount > 1000)
>   .map(_.amount){noformat}
> Or alternatively you could do this
> {noformat}
> dataStream.flatMap(_ match {
>   case p: Purchase if p.amount > 1000 => Some(p.amount)
>   case _ => None
> }){noformat}
> But with collect implemented it could look like
> {noformat}
> dataStream.collect {
>   case p: Purchase if p.amount > 1000 => p.amount
> }{noformat}
>  
> Which is a lot nicer to both read and write



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-8828) Add collect method to DataStream / DataSet scala api

2018-03-01 Thread Jelmer Kuperus (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-8828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jelmer Kuperus updated FLINK-8828:
--
Description: 
A collect function is a method that takes a Partial Function as its parameter 
and applies it to all the elements in the collection to create a new collection 
which satisfies the Partial Function.

It can be found on all [core scala collection 
classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
 as well as on spark's [rdd 
interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]

To understand its utility imagine the following scenario :

You have a DataStream that produces events of type _Purchase_ and _View_ 
 You would like to transform this stream into a stream of purchase amounts over 
1000 euros.

Currently an implementation might look like
{noformat}
val x = dataStream
  .filter(_.isInstanceOf[Purchase])
  .map(_.asInstanceOf[Purchase])
  .filter(_.amount > 1000)
  .map(_.amount){noformat}
Or alternatively you could do this
{noformat}
dataStream.flatMap(_ match {
  case p: Purchase if p.amount > 1000 => Some(p.amount)
  case _ => None
}){noformat}
But with collect implemented it could look like
{noformat}
dataStream.collect {
  case p: Purchase if p.amount > 1000 => p.amount
}{noformat}
 

Which is a lot nicer to both read and write

  was:
A collect function is a method that takes a Partial Function as its parameter 
and applies it to all the elements in the collection to create a new collection 
which satisfies the Partial Function.

It can be found on all [core scala collection 
classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
 as well as on spark's [rdd 
interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]

To understand its utility imagine the following scenario :

You have a DataStream that produces events of type _Purchase_ and View 
 You would like to transform this stream into a stream of purchase amounts over 
1000 euros.

Currently an implementation might look like
{noformat}
val x = dataStream
  .filter(_.isInstanceOf[Purchase])
  .map(_.asInstanceOf[Purchase])
  .filter(_.amount > 1000)
  .map(_.amount){noformat}
Or alternatively you could do this
{noformat}
dataStream.flatMap(_ match {
  case p: Purchase if p.amount > 1000 => Some(p.amount)
  case _ => None
}){noformat}
But with collect implemented it could look like
{noformat}
dataStream.collect {
  case p: Purchase if p.amount > 1000 => p.amount
}{noformat}
 

Which is a lot nicer to both read and write


> Add collect method to DataStream / DataSet scala api
> 
>
> Key: FLINK-8828
> URL: https://issues.apache.org/jira/browse/FLINK-8828
> Project: Flink
>  Issue Type: Improvement
>  Components: Core, DataSet API, DataStream API, Scala API
>Affects Versions: 1.4.0
>Reporter: Jelmer Kuperus
>Priority: Major
>
> A collect function is a method that takes a Partial Function as its parameter 
> and applies it to all the elements in the collection to create a new 
> collection which satisfies the Partial Function.
> It can be found on all [core scala collection 
> classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
>  as well as on spark's [rdd 
> interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]
> To understand its utility imagine the following scenario :
> You have a DataStream that produces events of type _Purchase_ and _View_ 
>  You would like to transform this stream into a stream of purchase amounts 
> over 1000 euros.
> Currently an implementation might look like
> {noformat}
> val x = dataStream
>   .filter(_.isInstanceOf[Purchase])
>   .map(_.asInstanceOf[Purchase])
>   .filter(_.amount > 1000)
>   .map(_.amount){noformat}
> Or alternatively you could do this
> {noformat}
> dataStream.flatMap(_ match {
>   case p: Purchase if p.amount > 1000 => Some(p.amount)
>   case _ => None
> }){noformat}
> But with collect implemented it could look like
> {noformat}
> dataStream.collect {
>   case p: Purchase if p.amount > 1000 => p.amount
> }{noformat}
>  
> Which is a lot nicer to both read and write



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Updated] (FLINK-8828) Add collect method to DataStream / DataSet scala api

2018-03-01 Thread Jelmer Kuperus (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-8828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jelmer Kuperus updated FLINK-8828:
--
Description: 
A collect function is a method that takes a Partial Function as its parameter 
and applies it to all the elements in the collection to create a new collection 
which satisfies the Partial Function.

It can be found on all [core scala collection 
classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
 as well as on spark's [rdd 
interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]

To understand its utility imagine the following scenario :

You have a DataStream that produces events of type _Purchase_ and View 
 You would like to transform this stream into a stream of purchase amounts over 
1000 euros.

Currently an implementation might look like
{noformat}
val x = dataStream
  .filter(_.isInstanceOf[Purchase])
  .map(_.asInstanceOf[Purchase])
  .filter(_.amount > 1000)
  .map(_.amount){noformat}
Or alternatively you could do this
{noformat}
dataStream.flatMap(_ match {
  case p: Purchase if p.amount > 1000 => Some(p.amount)
  case _ => None
}){noformat}
But with collect implemented it could look like
{noformat}
dataStream.collect {
  case p: Purchase if p.amount > 1000 => p.amount
}{noformat}
 

Which is a lot nicer to both read and write

  was:
A collect function is a method that takes a Partial Function as its parameter 
and applies it to all the elements in the collection to create a new collection 
which satisfies the Partial Function.

It can be found on all [core scala collection 
classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
 as well as on spark's [rdd 
interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]

To understand its utility imagine the following scenario :(

You have a DataStream that produces events of type _Purchase_ and View 
You would like to transform this stream into a stream of purchase amounts over 
1000 euros. 

Currently an implementation might look like
{noformat}
val x = dataStream
  .filter(_.isInstanceOf[Purchase])
  .map(_.asInstanceOf[Purchase])
  .filter(_.amount > 1000)
  .map(_.amount){noformat}
Or alternatively you could do this
{noformat}
dataStream.flatMap(_ match {
  case p: Purchase if p.amount > 1000 => Some(p.amount)
  case _ => None
}){noformat}
But with collect implemented it could look like
{noformat}
dataStream.collect {
  case p: Purchase if p.amount > 1000 => p.amount
}{noformat}
 

Which is a lot nicer to both read and write


> Add collect method to DataStream / DataSet scala api
> 
>
> Key: FLINK-8828
> URL: https://issues.apache.org/jira/browse/FLINK-8828
> Project: Flink
>  Issue Type: Improvement
>  Components: Core, DataSet API, DataStream API, Scala API
>Affects Versions: 1.4.0
>Reporter: Jelmer Kuperus
>Priority: Major
>
> A collect function is a method that takes a Partial Function as its parameter 
> and applies it to all the elements in the collection to create a new 
> collection which satisfies the Partial Function.
> It can be found on all [core scala collection 
> classes|http://www.scala-lang.org/api/2.9.2/scala/collection/TraversableLike.html]
>  as well as on spark's [rdd 
> interface|https://spark.apache.org/docs/2.2.0/api/scala/index.html#org.apache.spark.rdd.RDD]
> To understand its utility imagine the following scenario :
> You have a DataStream that produces events of type _Purchase_ and View 
>  You would like to transform this stream into a stream of purchase amounts 
> over 1000 euros.
> Currently an implementation might look like
> {noformat}
> val x = dataStream
>   .filter(_.isInstanceOf[Purchase])
>   .map(_.asInstanceOf[Purchase])
>   .filter(_.amount > 1000)
>   .map(_.amount){noformat}
> Or alternatively you could do this
> {noformat}
> dataStream.flatMap(_ match {
>   case p: Purchase if p.amount > 1000 => Some(p.amount)
>   case _ => None
> }){noformat}
> But with collect implemented it could look like
> {noformat}
> dataStream.collect {
>   case p: Purchase if p.amount > 1000 => p.amount
> }{noformat}
>  
> Which is a lot nicer to both read and write



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)