GitHub user peterableda opened a pull request:
https://github.com/apache/spark/pull/16172
[SPARK-18740] Log spark.app.name in driver logs
## What changes were proposed in this pull request?
Added simple logInfo line to print out the `spark.app.name` in the driver
logs
GitHub user peterableda opened a pull request:
https://github.com/apache/spark/pull/13850
[SPARK-16138] Try to cancel executor requests only if we have at least 1
## What changes were proposed in this pull request?
Adding additional check to if statement
## How
Github user peterableda commented on a diff in the pull request:
https://github.com/apache/spark/pull/13552#discussion_r66394240
--- Diff:
yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocator.scala ---
@@ -353,7 +353,7 @@ private[yarn] class YarnAllocator
Github user peterableda commented on a diff in the pull request:
https://github.com/apache/spark/pull/13552#discussion_r66262459
--- Diff:
yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocator.scala ---
@@ -353,7 +353,7 @@ private[yarn] class YarnAllocator
Github user peterableda commented on a diff in the pull request:
https://github.com/apache/spark/pull/13552#discussion_r66251071
--- Diff:
yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocator.scala ---
@@ -353,7 +353,7 @@ private[yarn] class YarnAllocator
Github user peterableda commented on a diff in the pull request:
https://github.com/apache/spark/pull/13552#discussion_r66220706
--- Diff:
yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocator.scala ---
@@ -353,7 +353,7 @@ private[yarn] class YarnAllocator
Github user peterableda commented on a diff in the pull request:
https://github.com/apache/spark/pull/13552#discussion_r66217380
--- Diff:
yarn/src/main/scala/org/apache/spark/deploy/yarn/YarnAllocator.scala ---
@@ -353,7 +353,7 @@ private[yarn] class YarnAllocator
GitHub user peterableda opened a pull request:
https://github.com/apache/spark/pull/13552
[SPARK-15813] Use past tense for the cancel container request message
## What changes were proposed in this pull request?
Use past tense for the cancel container request message
Github user peterableda commented on the pull request:
https://github.com/apache/spark/pull/12395#issuecomment-213471409
@srowen OK, how about reverting my last commits and setting a 32MB minimum
for the legacy MemoryManager - keeping the two implementation separate?
---
If your
Github user peterableda commented on the pull request:
https://github.com/apache/spark/pull/12395#issuecomment-212418413
@srowen I refactored the checks based on your recommendations. Please let
me know if you see something go against our conventions or you would do
differently
Github user peterableda commented on a diff in the pull request:
https://github.com/apache/spark/pull/12395#discussion_r59849405
--- Diff:
core/src/main/scala/org/apache/spark/memory/StaticMemoryManager.scala ---
@@ -104,6 +104,8 @@ private[spark] class StaticMemoryManager
Github user peterableda commented on a diff in the pull request:
https://github.com/apache/spark/pull/12395#discussion_r59743938
--- Diff:
core/src/main/scala/org/apache/spark/memory/StaticMemoryManager.scala ---
@@ -104,6 +104,8 @@ private[spark] class StaticMemoryManager
GitHub user peterableda opened a pull request:
https://github.com/apache/spark/pull/12395
Add minimum memory checks for drivers and executors
## What changes were proposed in this pull request?
Implement the same memory size validations for the StaticMemoryManager
(Legacy
GitHub user peterableda opened a pull request:
https://github.com/apache/spark/pull/12392
[SPARK-14633] Use more readable format to show memory bytes in Error Message
## What changes were proposed in this pull request?
Round memory bytes and convert it to Long to itâs
Github user peterableda closed the pull request at:
https://github.com/apache/spark/pull/11312
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature
Github user peterableda commented on the pull request:
https://github.com/apache/spark/pull/11312#issuecomment-189749198
You are right, the timestamp here refers to the processing time of the
batch. I don't have a use-case currently where I want to capture this
information
Github user peterableda commented on the pull request:
https://github.com/apache/spark/pull/11312#issuecomment-189619833
Conceptually this is how Hive works:
Create a simple Hive table definition for a `test` table for the location
`/data/`
```
CREATE EXTERNAL TABLE
Github user peterableda commented on the pull request:
https://github.com/apache/spark/pull/11312#issuecomment-189236579
@srowen The idea was to create an external Hive table on top of the output
of the Spark streaming application.
---
If your project is set up for it, you can
GitHub user peterableda opened a pull request:
https://github.com/apache/spark/pull/11312
[SPARK-13438][STREAMING] Remove by default dash from filename
Spark generates the following schema with prefix = `/data/timestamp=`
```
/data/timestamp=-1455894364000/_SUCCESS
/data
19 matches
Mail list logo