Repository: spark
Updated Branches:
refs/heads/branch-2.1 3c8861d92 -> 162bdb910
[SPARK-18031][TESTS] Fix flaky test ExecutorAllocationManagerSuite.basic
functionality
## What changes were proposed in this pull request?
The failure is because in `test("basic functionality")`, it doesn't
Repository: spark
Updated Branches:
refs/heads/master 354e93618 -> 95efc895e
[SPARK-18588][SS][KAFKA] Create a new KafkaConsumer when error happens to fix
the flaky test
## What changes were proposed in this pull request?
When KafkaSource fails on Kafka errors, we should create a new
Repository: spark
Updated Branches:
refs/heads/branch-2.1 021952d58 -> 9a3c5bd70
[FLAKY-TEST] InputStreamsSuite.socket input stream
## What changes were proposed in this pull request?
Repository: spark
Updated Branches:
refs/heads/master 7e8994ffd -> afe36516e
[FLAKY-TEST] InputStreamsSuite.socket input stream
## What changes were proposed in this pull request?
ot;
- Changed package of InternalOutputModes from o.a.s.sql to o.a.s.sql.catalyst
- Added update mode state removing with watermark to StateStoreSaveExec
## How was this patch tested?
Added new tests in changed modules
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #16360 fr
ot;
- Changed package of InternalOutputModes from o.a.s.sql to o.a.s.sql.catalyst
- Added update mode state removing with watermark to StateStoreSaveExec
## How was this patch tested?
Added new tests in changed modules
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #16360 fr
Repository: spark
Updated Branches:
refs/heads/branch-2.1 0e51bb085 -> 17ef57fe8
[SPARK-18588][SS][KAFKA] Create a new KafkaConsumer when error happens to fix
the flaky test
## What changes were proposed in this pull request?
When KafkaSource fails on Kafka errors, we should create a new
ean that we should fall back to what's
in the offset log.
- A OneTime trigger execution that results in an exception being thrown.
marmbrus tdas zsxwing
Please review http://spark.apache.org/contributing.html before opening a pull
request.
Author: Tyson Condie <tcon...@gmail.com>
Author:
Repository: spark
Updated Branches:
refs/heads/master 80fd07038 -> 82b598b96
http://git-wip-us.apache.org/repos/asf/spark/blob/82b598b9/sql/core/src/test/scala/org/apache/spark/sql/streaming/FlatMapGroupsWithStateSuite.scala
is would make it more
general if you extends this operation to RelationGroupedDataset and python APIs.
## How was this patch tested?
Existing unit tests.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #17385 from tdas/SPARK-20057.
Project: http://git-wip-us.apache.org/repos/a
il.com>
Closes #17488 from tdas/SPARK-20165.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/567a50ac
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/567a50ac
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff
to
address.
## How was this patch tested?
New unit tests in
- MapGroupsWithStateSuite for timeouts.
- StateStoreSuite for new APIs in StateStore.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #17179 from tdas/mapgroupwithstate-timeout.
Project: http://git-wip-us.apache.org
Repository: spark
Updated Branches:
refs/heads/master 0ee9fbf51 -> 990af630d
http://git-wip-us.apache.org/repos/asf/spark/blob/990af630/sql/core/src/main/scala/org/apache/spark/sql/streaming/KeyedState.scala
--
diff --git
Repository: spark
Updated Branches:
refs/heads/master 7b5d873ae -> 376d78216
[SPARK-19986][TESTS] Make pyspark.streaming.tests.CheckpointTests more stable
## What changes were proposed in this pull request?
Sometimes, CheckpointTests will hang on a busy machine because the streaming
jobs
Repository: spark
Updated Branches:
refs/heads/branch-2.1 710b5554e -> 5fb70831b
[SPARK-19986][TESTS] Make pyspark.streaming.tests.CheckpointTests more stable
## What changes were proposed in this pull request?
Sometimes, CheckpointTests will hang on a busy machine because the streaming
Repository: spark
Updated Branches:
refs/heads/master 7de66bae5 -> 3783539d7
[SPARK-19873][SS] Record num shuffle partitions in offset log and enforce in
next batch.
## What changes were proposed in this pull request?
If the user changes the shuffle partition number between batches,
fka.
zsxwing tdas
Please review http://spark.apache.org/contributing.html before opening a pull
request.
Author: Tyson Condie <tcon...@gmail.com>
Closes #17246 from tcondie/kafka-write-docs.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/
Repository: spark
Updated Branches:
refs/heads/branch-2.0 fd5149aff -> 6ee7d5bf4
[SPARK-19986][TESTS] Make pyspark.streaming.tests.CheckpointTests more stable
## What changes were proposed in this pull request?
Sometimes, CheckpointTests will hang on a busy machine because the streaming
Repository: spark
Updated Branches:
refs/heads/master 9281a3d50 -> 2d73fcced
[SPARK-20051][SS] Fix StreamSuite flaky test - recover from v2.1 checkpoint
## What changes were proposed in this pull request?
There is a race condition between calling stop on a streaming query and
deleting
ing `KeyedState.setTimeoutTimestamp`. The keys times out when the watermark
crosses the timeout timestamp.
## How was this patch tested?
Unit tests
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #17361 from tdas/SPARK-20030.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Com
Repository: spark
Updated Branches:
refs/heads/master 99a947312 -> 924c42477
[SPARK-20301][FLAKY-TEST] Fix Hadoop Shell.runCommand flakiness in Structured
Streaming tests
## What changes were proposed in this pull request?
Some Structured Streaming tests show flakiness such as:
```
[info] -
Repository: spark
Updated Branches:
refs/heads/master e9e2c612d -> 1bf901238
[SPARK-19858][SS] Add output mode to flatMapGroupsWithState and disallow
invalid cases
## What changes were proposed in this pull request?
Add a output mode parameter to `flatMapGroupsWithState` and just define
end and ErrorIfExist supported under identical semantics. Other
save modes result in an AnalysisException
tdas zsxwing
## How was this patch tested?
### The following unit tests will be included
- write to stream with topic field: valid stream write with data that includes
an existing topic in the sch
end and ErrorIfExist supported under identical semantics. Other
save modes result in an AnalysisException
tdas zsxwing
## How was this patch tested?
### The following unit tests will be included
- write to stream with topic field: valid stream write with data that includes
an existing topic in the schema
- wr
ple,
ProcessingTimeExecutorSuite does not need to create any context for testing,
just needs the StreamManualClock.
## How was this patch tested?
Added new unit tests to comprehensively test this behavior.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #17525 from tdas/SPARK-20209.
Project: http:
ted markdown docs
- Updated scala docs
- Added scala and Java example
## How was this patch tested?
Manually ran examples.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #17539 from tdas/SPARK-20224.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http:
ate when using timeouts.
## How was this patch tested?
manually ran the example
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #17676 from tdas/SPARK-20377.
(cherry picked from commit 74aa0df8f7f132b62754e5159262e4a5b9b641ab)
Signed-off-by: Tathagata Das <tathagata.das1.
hen using timeouts.
## How was this patch tested?
manually ran the example
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #17676 from tdas/SPARK-20377.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/74aa
Repository: spark
Updated Branches:
refs/heads/master fbc269252 -> 282f00b41
[SPARK-21696][SS] Fix a potential issue that may generate partial snapshot files
## What changes were proposed in this pull request?
Directly writing a snapshot file may generate a partial file. This PR changes
it
Repository: spark
Updated Branches:
refs/heads/branch-2.2 7b9807754 -> 48bacd36c
[SPARK-21696][SS] Fix a potential issue that may generate partial snapshot files
## What changes were proposed in this pull request?
Directly writing a snapshot file may generate a partial file. This PR changes
Repository: spark
Updated Branches:
refs/heads/master e16e8c7ad -> e0af76a36
[SPARK-21370][SS] Add test for state reliability when one read-only state store
aborts after read-write state store commits
## What changes were proposed in this pull request?
During Streaming Aggregation, we have
PR enables
that.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #18661 from tdas/SPARK-21409-2.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/e9faae13
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree
lso, removed recently added numPartitions from StatefulOperatorProgress as
this value does not change through the query run, and there are other ways to
find that.
## How was this patch tested?
Updated unit tests
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #18675 from t
ing by
removing its uses from tests as much as possible.
## How was this patch tested?
Existing tests.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #18678 from tdas/SPARK-21464.
(cherry picked from commit 70fe99dc62ef636a99bcb8a580ad4de4dca95181)
Signed-off-by: Tathagata Das <ta
ing its uses from tests as much as possible.
## How was this patch tested?
Existing tests.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #18678 from tdas/SPARK-21464.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spa
row interrupt exception, in
which case temporary checkpoint directories will not be deleted, and the test
will fail.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #18442 from tdas/DatastreamReaderWriterSuite-fix.
(cherry picked from commit 60043f22458668ac7ecba94fa78953f23a6bdcec)
S
upt exception, in
which case temporary checkpoint directories will not be deleted, and the test
will fail.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #18442 from tdas/DatastreamReaderWriterSuite-fix.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http:
Repository: spark
Updated Branches:
refs/heads/master 606432a13 -> 01c999e7f
[SPARK-20461][CORE][SS] Use UninterruptibleThread for Executor and fix the
potential hang in CachedKafkaConsumer
## What changes were proposed in this pull request?
This PR changes Executor's threads to
Repository: spark
Updated Branches:
refs/heads/branch-2.2 4512e2ae6 -> 753e129f3
[SPARK-20461][CORE][SS] Use UninterruptibleThread for Executor and fix the
potential hang in CachedKafkaConsumer
## What changes were proposed in this pull request?
This PR changes Executor's threads to
Repository: spark
Updated Branches:
refs/heads/branch-2.2 753e129f3 -> 3d53d825e
[SPARK-20452][SS][KAFKA] Fix a potential ConcurrentModificationException for
batch Kafka DataFrame
## What changes were proposed in this pull request?
Cancel a batch Kafka query but one of task cannot be
Repository: spark
Updated Branches:
refs/heads/master 01c999e7f -> 823baca2c
[SPARK-20452][SS][KAFKA] Fix a potential ConcurrentModificationException for
batch Kafka DataFrame
## What changes were proposed in this pull request?
Cancel a batch Kafka query but one of task cannot be cancelled,
Repository: spark
Updated Branches:
refs/heads/master fb54a564d -> 6edfff055
[SPARK-21596][SS] Ensure places calling HDFSMetadataLog.get check the return
value
## What changes were proposed in this pull request?
When I was investigating a flaky test, I realized that many places don't check
Repository: spark
Updated Branches:
refs/heads/master 2d799d080 -> 0fb73253f
[SPARK-21587][SS] Added filter pushdown through watermarks.
## What changes were proposed in this pull request?
Push filter predicates through EventTimeWatermark if they're deterministic and
do not reference the
il.com>
Closes #18355 from tdas/SPARK-21145.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/fe24634d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/fe24634d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff
Repository: spark
Updated Branches:
refs/heads/master 06c054411 -> bc537e40a
[SPARK-20957][SS][TESTS] Fix o.a.s.sql.streaming.StreamingQueryManagerSuite
listing
## What changes were proposed in this pull request?
When stopping StreamingQuery, StreamExecution will set `streamDeathCause` then
Repository: spark
Updated Branches:
refs/heads/branch-2.2 1388fdd70 -> 421d8ecb8
[SPARK-20957][SS][TESTS] Fix o.a.s.sql.streaming.StreamingQueryManagerSuite
listing
## What changes were proposed in this pull request?
When stopping StreamingQuery, StreamExecution will set `streamDeathCause`
ing the metadata of top-level aliases.
## How was this patch tested?
New unit test
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #19240 from tdas/SPARK-22018.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8866
Repository: spark
Updated Branches:
refs/heads/master c7307acda -> 0bad10d3e
[SPARK-22017] Take minimum of all watermark execs in StreamExecution.
## What changes were proposed in this pull request?
Take the minimum of all watermark exec nodes as the "real" watermark in
StreamExecution,
Repository: spark
Updated Branches:
refs/heads/master a8a5cd24e -> f32a84250
http://git-wip-us.apache.org/repos/asf/spark/blob/f32a8425/sql/core/src/test/scala/org/apache/spark/sql/streaming/StateStoreMetricsTest.scala
--
diff
evented stream-stream join on an empty batch dataframe to be collapsed by
the optimizer
## How was this patch tested?
- New tests in StreamingJoinSuite
- Updated tests UnsupportedOperationSuite
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #19271 from tdas/SPARK-22053.
Proje
ll, and avoid these
confusing corner cases.
## How was this patch tested?
Refactored tests.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #19416 from tdas/SPARK-22187.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/sp
Repository: spark
Updated Branches:
refs/heads/master 014dc8471 -> e8547ffb4
[SPARK-22238] Fix plan resolution bug caused by EnsureStatefulOpPartitioning
## What changes were proposed in this pull request?
In EnsureStatefulOpPartitioning, we check that the inputRDD to a SparkPlan has
the
Repository: spark
Updated Branches:
refs/heads/branch-2.2 fb1b5f08a -> 1f7c4869b
[SPARK-21925] Update trigger interval documentation in docs with behavior
change in Spark 2.2
Forgot to update docs with behavior change.
Author: Burak Yavuz
Closes #19138 from
Repository: spark
Updated Branches:
refs/heads/master 2974406d1 -> 8c954d2cd
[SPARK-21925] Update trigger interval documentation in docs with behavior
change in Spark 2.2
Forgot to update docs with behavior change.
Author: Burak Yavuz
Closes #19138 from
Repository: spark
Updated Branches:
refs/heads/master 36b48ee6e -> acdf45fb5
[SPARK-21765] Check that optimization doesn't affect isStreaming bit.
## What changes were proposed in this pull request?
Add an assert in logical plan optimization that the isStreaming bit stays the
same, and fix
Repository: spark
Updated Branches:
refs/heads/master 2dd37d827 -> d3abb3699
[SPARK-21788][SS] Handle more exceptions when stopping a streaming query
## What changes were proposed in this pull request?
Add more cases we should view as a normal query stop rather than a failure.
## How was
[SPARK-22136][SS] Implement stream-stream outer joins.
## What changes were proposed in this pull request?
Allow one-sided outer joins between two streams when a watermark is defined.
## How was this patch tested?
new unit tests
Author: Jose Torres
Closes #19327 from
Repository: spark
Updated Branches:
refs/heads/master 5f6943345 -> 3099c574c
http://git-wip-us.apache.org/repos/asf/spark/blob/3099c574/sql/core/src/test/scala/org/apache/spark/sql/streaming/StreamingJoinSuite.scala
--
diff
Repository: spark
Updated Branches:
refs/heads/master e1960c3d6 -> 75d666b95
[SPARK-22136][SS] Evaluate one-sided conditions early in stream-stream joins.
## What changes were proposed in this pull request?
Evaluate one-sided conditions early in stream-stream joins.
This is in addition to
Das <tathagata.das1...@gmail.com>
Closes #19495 from tdas/SPARK-22278.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/f3137fee
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/f3137fee
Diff: http://git-wip-us.apache.org/repos/
ation
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #21220 from tdas/SPARK-24157.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/47b5b685
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/47b5b685
D
Repository: spark
Updated Branches:
refs/heads/master d04806a23 -> af4dc5028
[SPARK-24039][SS] Do continuous processing writes with multiple compute() calls
## What changes were proposed in this pull request?
Do continuous processing writes with multiple compute() calls.
The current
Repository: spark
Updated Branches:
refs/heads/master 03e90f65b -> a33dcf4a0
[SPARK-24234][SS] Reader for continuous processing shuffle
## What changes were proposed in this pull request?
Read RDD for continuous processing shuffle, as well as the initial RPC-based
row receiver.
Repository: spark
Updated Branches:
refs/heads/master b7a036b75 -> f45793329
[SPARK-23416][SS] Add a specific stop method for ContinuousExecution.
## What changes were proposed in this pull request?
Add a specific stop method for ContinuousExecution. The previous
StreamExecution.stop()
Repository: spark
Updated Branches:
refs/heads/master 53c06ddab -> 0fd68cb72
[SPARK-24234][SS] Support multiple row writers in continuous processing shuffle
reader.
## What changes were proposed in this pull request?
nup. This PR enables it for stream-stream joins.
## How was this patch tested?
- Updated join tests. Additionally, updated them to not use `CheckLastBatch`
anywhere to set good precedence for future.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #21253 from tdas/SPARK-24158.
Repository: spark
Updated Branches:
refs/heads/master 710e4e81a -> 434d74e33
[SPARK-23503][SS] Enforce sequencing of committed epochs for Continuous
Execution
## What changes were proposed in this pull request?
Made changes to EpochCoordinator so that it enforces a commit order. In case a
Das
Closes #21477 from tdas/SPARK-24396.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b5ccf0d3
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b5ccf0d3
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b5ccf
ext. It mirrors the Java TaskContext API of returning a string value if
the key exists, or None if the key does not exist.
## How was this patch tested?
New test added.
Author: Tathagata Das
Closes #21437 from tdas/SPARK-24397.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: h
lag appropriately.
## How was this patch tested?
new unit test
Author: Tathagata Das
Closes #21491 from tdas/SPARK-24453.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2c2a86b5
Tree: http://git-wip-us.apache.org/repos/
ion is running. Great for
debugging production issues.
## How was this patch tested?
Not necessary.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #21160 from tdas/SPARK-24094.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/
Repository: spark
Updated Branches:
refs/heads/branch-2.3 be5991902 -> 44763d93c
[SPARK-22912] v2 data source support in MicroBatchExecution
## What changes were proposed in this pull request?
Support for v2 data sources in microbatch streaming.
## How was this patch tested?
A very basic
Repository: spark
Updated Branches:
refs/heads/master eed82a0b2 -> 4f7e75883
[SPARK-22912] v2 data source support in MicroBatchExecution
## What changes were proposed in this pull request?
Support for v2 data sources in microbatch streaming.
## How was this patch tested?
A very basic new
How was this patch tested?
new unit test
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20311 from tdas/SPARK-23144.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/bf34d665
Tree: http://git-wip-us.apache.org/repos/
How was this patch tested?
new unit test
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20311 from tdas/SPARK-23144.
(cherry picked from commit bf34d665b9c865e00fac7001500bf6d521c2dff9)
Signed-off-by: Tathagata Das <tathagata.das1...@gmail.com>
Project: http://git-wip-us.a
Das <tathagata.das1...@gmail.com>
Closes #20309 from tdas/SPARK-23143.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2d41f040
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/2d41f040
Diff: http://git-wip-us.a
[SPARK-22908][SS] Roll forward continuous processing Kafka support with fix to
continuous Kafka data reader
## What changes were proposed in this pull request?
The Kafka reader is now interruptible and can close itself.
## How was this patch tested?
I locally ran one of the
Repository: spark
Updated Branches:
refs/heads/branch-2.3 08252bb38 -> 0a441d2ed
http://git-wip-us.apache.org/repos/asf/spark/blob/0a441d2e/external/kafka-0-10-sql/src/test/scala/org/apache/spark/sql/kafka010/KafkaSourceSuite.scala
[SPARK-22908][SS] Roll forward continuous processing Kafka support with fix to
continuous Kafka data reader
## What changes were proposed in this pull request?
The Kafka reader is now interruptible and can close itself.
## How was this patch tested?
I locally ran one of the
Repository: spark
Updated Branches:
refs/heads/master a9b845ebb -> 166705785
http://git-wip-us.apache.org/repos/asf/spark/blob/16670578/external/kafka-0-10-sql/src/test/scala/org/apache/spark/sql/kafka010/KafkaSourceSuite.scala
Repository: spark
Updated Branches:
refs/heads/master c132538a1 -> 86a845031
[SPARK-23033][SS] Don't use task level retry for continuous processing
## What changes were proposed in this pull request?
Continuous processing tasks will fail on any attempt number greater than 0.
Repository: spark
Updated Branches:
refs/heads/branch-2.3 1a6dfaf25 -> dbd2a5566
[SPARK-23033][SS] Don't use task level retry for continuous processing
## What changes were proposed in this pull request?
Continuous processing tasks will fail on any attempt number greater than 0.
Repository: spark
Updated Branches:
refs/heads/branch-2.3 3a80cc59b -> 2a87c3a77
[SPARK-23052][SS] Migrate ConsoleSink to data source V2 api.
## What changes were proposed in this pull request?
Migrate ConsoleSink to data source V2 api.
Note that this includes a missing piece in
Repository: spark
Updated Branches:
refs/heads/master 39d244d92 -> 1c76a91e5
[SPARK-23052][SS] Migrate ConsoleSink to data source V2 api.
## What changes were proposed in this pull request?
Migrate ConsoleSink to data source V2 api.
Note that this includes a missing piece in
Repository: spark
Updated Branches:
refs/heads/branch-2.3 b94debd2b -> f891ee324
http://git-wip-us.apache.org/repos/asf/spark/blob/f891ee32/sql/core/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala
--
diff --git
Repository: spark
Updated Branches:
refs/heads/master 0b2eefb67 -> 6f7aaed80
http://git-wip-us.apache.org/repos/asf/spark/blob/6f7aaed8/sql/core/src/main/scala/org/apache/spark/sql/DataFrameWriter.scala
--
diff --git
[SPARK-22908] Add kafka source and sink for continuous processing.
## What changes were proposed in this pull request?
Add kafka source and sink for continuous processing. This involves two small
changes to the execution engine:
* Bring data reader close() into the normal data reader thread to
[SPARK-22908] Add kafka source and sink for continuous processing.
## What changes were proposed in this pull request?
Add kafka source and sink for continuous processing. This involves two small
changes to the execution engine:
* Bring data reader close() into the normal data reader thread to
How was this patch tested?
N/A
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20494 from tdas/SPARK-23064-2.
(cherry picked from commit eaf35de2471fac4337dd2920026836d52b1ec847)
Signed-off-by: Tathagata Das <tathagata.das1...@gmail.com>
Project: http://git-wip-us.apache.org/repos
tch tested?
N/A
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20494 from tdas/SPARK-23064-2.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/eaf35de2
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree
sts.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Author: Burak Yavuz <brk...@gmail.com>
Closes #20445 from tdas/SPARK-23092.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/30295bf5
Tree: http://git-wip
key#9, value#12]
+- Project [value#66 AS value#12]// solution: project with aliases
+- LocalRelation [value#66]
```
## How was this patch tested?
New unit test
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20598 from tdas/SPARK-23406.
Project: http://git
ove this)
Please review http://spark.apache.org/contributing.html before opening a pull
request.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20631 from tdas/SPARK-23454.
(cherry picked from commit 601d653bff9160db8477f86d961e609fc2190237)
Signed-off-by: Tathagata Das <ta
ove this)
Please review http://spark.apache.org/contributing.html before opening a pull
request.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20631 from tdas/SPARK-23454.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spa
rom multiple threads - the query thread at the time of reader factory
creation, and the epoch tracking thread at the time of `needsReconfiguration`.
## How was this patch tested?
Existing tests.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20655 from tdas/SPARK-23484.
rom multiple threads - the query thread at the time of reader factory
creation, and the epoch tracking thread at the time of `needsReconfiguration`.
## How was this patch tested?
Existing tests.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20655 from tdas/SPARK-23484.
Proj
throughput of V1
and V2 using 20M records in a single partition. They were comparable.
## How was this patch tested?
Existing tests, few modified to be better tests than the existing ones.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20554 from tdas/SPARK-23362.
Project: http://g
Repository: spark
Updated Branches:
refs/heads/master c5857e496 -> 0a73aa31f
http://git-wip-us.apache.org/repos/asf/spark/blob/0a73aa31/external/kafka-0-10-sql/src/test/scala/org/apache/spark/sql/kafka010/KafkaSourceSuite.scala
ing sure that the flaky tests are deterministic.
## How was this patch tested?
Modified test cases in `Streaming*JoinSuites` where there are consecutive
`AddData` actions.
Author: Tathagata Das <tathagata.das1...@gmail.com>
Closes #20650 from tdas/SPARK-23408.
Project: http:
Repository: spark
Updated Branches:
refs/heads/master 185f5bc7d -> 7ec83658f
[SPARK-23491][SS] Remove explicit job cancellation from ContinuousExecution
reconfiguring
## What changes were proposed in this pull request?
Remove queryExecutionThread.interrupt() from ContinuousExecution. As
701 - 800 of 847 matches
Mail list logo