+1

Yuming Wang <yumw...@apache.org> 于2025年5月29日周四 02:22写道:

> +1.
>
> On Thu, May 29, 2025 at 3:36 PM DB Tsai <dbt...@dbtsai.com> wrote:
>
>> +1
>> Sent from my iPhone
>>
>> On May 29, 2025, at 12:15 AM, John Zhuge <jzh...@apache.org> wrote:
>>
>> 
>> +1 Nice feature
>>
>> On Wed, May 28, 2025 at 9:53 PM Yuanjian Li <xyliyuanj...@gmail.com>
>> wrote:
>>
>>> +1
>>>
>>> Kent Yao <y...@apache.org> 于2025年5月28日周三 19:31写道:
>>>
>>>> +1, LGTM.
>>>>
>>>> Kent
>>>>
>>>> 在 2025年5月29日星期四,Chao Sun <sunc...@apache.org> 写道:
>>>>
>>>>> +1. Super excited by this initiative!
>>>>>
>>>>> On Wed, May 28, 2025 at 1:54 PM Yanbo Liang <yblia...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> +1
>>>>>>
>>>>>> On Wed, May 28, 2025 at 12:34 PM huaxin gao <huaxin.ga...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> +1
>>>>>>> By unifying batch and low-latency streaming in Spark, we can
>>>>>>> eliminate the need for separate streaming engines, reducing system
>>>>>>> complexity and operational cost. Excited to see this direction!
>>>>>>>
>>>>>>> On Wed, May 28, 2025 at 9:08 AM Mich Talebzadeh <
>>>>>>> mich.talebza...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> My point about "in real time application or data, there is nothing
>>>>>>>> as an answer which is supposed to be late and correct. The timeliness 
>>>>>>>> is
>>>>>>>> part of the application. if I get the right answer too slowly it 
>>>>>>>> becomes
>>>>>>>> useless or wrong" is actually fundamental to *why* we need this
>>>>>>>> Spark Structured Streaming proposal.
>>>>>>>>
>>>>>>>> The proposal is precisely about enabling Spark to power
>>>>>>>> applications where, as I define it, the *timeliness* of the answer
>>>>>>>> is as critical as its *correctness*. Spark's current streaming
>>>>>>>> engine, primarily operating on micro-batches, often delivers results 
>>>>>>>> that
>>>>>>>> are technically "correct" but arrive too late to be truly useful for
>>>>>>>> certain high-stakes, real-time scenarios. This makes them "useless or
>>>>>>>> wrong" in a practical, business-critical sense.
>>>>>>>>
>>>>>>>> For example *in real-time fraud detection* and In *high-frequency
>>>>>>>> trading,* market data or trade execution commands must be
>>>>>>>> delivered with minimal latency. Even a slight delay can mean missed
>>>>>>>> opportunities or significant financial losses, making a "correct" price
>>>>>>>> update useless if it's not instantaneous. able for these demanding
>>>>>>>> use cases, where a "late but correct" answer is simply not good 
>>>>>>>> enough. As
>>>>>>>> a colliery it is a fundamental concept, so it has to be treated as 
>>>>>>>> such not
>>>>>>>> as a comment.in SPIP
>>>>>>>>
>>>>>>>> Hope this clarifies the connection in practical terms
>>>>>>>> Dr Mich Talebzadeh,
>>>>>>>> Architect | Data Science | Financial Crime | Forensic Analysis |
>>>>>>>> GDPR
>>>>>>>>
>>>>>>>>    view my Linkedin profile
>>>>>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Wed, 28 May 2025 at 16:32, Denny Lee <denny.g....@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Hey Mich,
>>>>>>>>>
>>>>>>>>> Sorry, I may be missing something here but what does your
>>>>>>>>> definition here have to do with the SPIP?   Perhaps add comments 
>>>>>>>>> directly
>>>>>>>>> to the SPIP to provide context as the code snippet below is a direct 
>>>>>>>>> copy
>>>>>>>>> from the SPIP itself.
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Denny
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Wed, May 28, 2025 at 06:48 Mich Talebzadeh <
>>>>>>>>> mich.talebza...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> just to add
>>>>>>>>>>
>>>>>>>>>> A stronger definition of real time. The engineering definition of
>>>>>>>>>> real time is roughly fast enough to be interactive
>>>>>>>>>>
>>>>>>>>>> However, I put a stronger definition. In real time application or
>>>>>>>>>> data, there is nothing as an answer which is supposed to be late and
>>>>>>>>>> correct. The timeliness is part of the application.if I get the right
>>>>>>>>>> answer too slowly it becomes useless or wrong
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Dr Mich Talebzadeh,
>>>>>>>>>> Architect | Data Science | Financial Crime | Forensic Analysis |
>>>>>>>>>> GDPR
>>>>>>>>>>
>>>>>>>>>>    view my Linkedin profile
>>>>>>>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Wed, 28 May 2025 at 11:10, Mich Talebzadeh <
>>>>>>>>>> mich.talebza...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> The current limitations in SSS come from micro-batching.If you
>>>>>>>>>>> are going to reduce micro-batching, this reduction must be balanced 
>>>>>>>>>>> against
>>>>>>>>>>> the available processing capacity of the cluster to prevent back 
>>>>>>>>>>> pressure
>>>>>>>>>>> and instability. In the case of Continuous Processing mode, a
>>>>>>>>>>> specific continuous trigger with a desired checkpoint interval quote
>>>>>>>>>>>
>>>>>>>>>>> "
>>>>>>>>>>> df.writeStream
>>>>>>>>>>>    .format("...")
>>>>>>>>>>>    .option("...")
>>>>>>>>>>>    .trigger(Trigger.RealTime(“300 Seconds”))    // new trigger
>>>>>>>>>>> type to enable real-time Mode
>>>>>>>>>>>    .start()
>>>>>>>>>>> This Trigger.RealTime signals that the query should run in the
>>>>>>>>>>> new ultra low-latency execution mode.  A time interval can also be
>>>>>>>>>>> specified, e.g. “300 Seconds”, to indicate how long each 
>>>>>>>>>>> micro-batch should
>>>>>>>>>>> run for.
>>>>>>>>>>> "
>>>>>>>>>>>
>>>>>>>>>>> will inevitably depend on many factors. Not that simple
>>>>>>>>>>> HTH
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Dr Mich Talebzadeh,
>>>>>>>>>>> Architect | Data Science | Financial Crime | Forensic Analysis |
>>>>>>>>>>> GDPR
>>>>>>>>>>>
>>>>>>>>>>>    view my Linkedin profile
>>>>>>>>>>> <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Wed, 28 May 2025 at 05:13, Jerry Peng <
>>>>>>>>>>> jerry.boyang.p...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi all,
>>>>>>>>>>>>
>>>>>>>>>>>> I want to start a discussion thread for the SPIP titled
>>>>>>>>>>>> “Real-Time Mode in Apache Spark Structured Streaming” that I've 
>>>>>>>>>>>> been
>>>>>>>>>>>> working on with Siying Dong, Indrajit Roy, Chao Sun, Jungtaek Lim, 
>>>>>>>>>>>> and
>>>>>>>>>>>> Michael Armbrust: [JIRA
>>>>>>>>>>>> <https://issues.apache.org/jira/browse/SPARK-52330>] [Doc
>>>>>>>>>>>> <https://docs.google.com/document/d/1CvJvtlTGP6TwQIT4kW6GFT1JbdziAYOBvt60ybb7Dw8/edit?usp=sharing>
>>>>>>>>>>>> ].
>>>>>>>>>>>>
>>>>>>>>>>>> The SPIP proposes a new execution mode called “Real-time Mode”
>>>>>>>>>>>> in Spark Structured Streaming that significantly lowers end-to-end 
>>>>>>>>>>>> latency
>>>>>>>>>>>> for processing streams of data.
>>>>>>>>>>>>
>>>>>>>>>>>> A key principle of this proposal is compatibility. Our goal is
>>>>>>>>>>>> to make Spark capable of handling streaming jobs that need results 
>>>>>>>>>>>> almost
>>>>>>>>>>>> immediately (within O(100) milliseconds). We want to achieve this 
>>>>>>>>>>>> without
>>>>>>>>>>>> changing the high-level DataFrame/Dataset API that users already 
>>>>>>>>>>>> use – so
>>>>>>>>>>>> existing streaming queries can run in this new ultra-low-latency 
>>>>>>>>>>>> mode by
>>>>>>>>>>>> simply turning it on, without rewriting their logic.
>>>>>>>>>>>>
>>>>>>>>>>>> In short, we’re trying to enable Spark to power real-time
>>>>>>>>>>>> applications (like instant anomaly alerts or live personalization) 
>>>>>>>>>>>> that
>>>>>>>>>>>> today cannot meet their latency requirements with Spark’s current 
>>>>>>>>>>>> streaming
>>>>>>>>>>>> engine.
>>>>>>>>>>>>
>>>>>>>>>>>> We'd greatly appreciate your feedback, thoughts, and
>>>>>>>>>>>> suggestions on this approach!
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best,
>>>>>> Yanbo
>>>>>>
>>>>>
>>
>> --
>> John Zhuge
>>
>>

Reply via email to