GitHub user orenmazor opened a pull request:
https://github.com/apache/spark/pull/4341
Sha jar staging
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/Shopify/spark sha_jar_staging
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/spark/pull/4341.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #4341
----
commit 007ae14b0f9feb9555655012ae5c60eccdb4c4a3
Author: Cheng Hao <[email protected]>
Date: 2014-12-12T06:51:49Z
[SPARK-4825] [SQL] CTAS fails to resolve when created using saveAsTable
Fix bug when query like:
```
test("save join to table") {
val testData = sparkContext.parallelize(1 to 10).map(i => TestData(i,
i.toString))
sql("CREATE TABLE test1 (key INT, value STRING)")
testData.insertInto("test1")
sql("CREATE TABLE test2 (key INT, value STRING)")
testData.insertInto("test2")
testData.insertInto("test2")
sql("SELECT COUNT(a.value) FROM test1 a JOIN test2 b ON a.key =
b.key").saveAsTable("test")
checkAnswer(
table("test"),
sql("SELECT COUNT(a.value) FROM test1 a JOIN test2 b ON a.key =
b.key").collect().toSeq)
}
```
Author: Cheng Hao <[email protected]>
Closes #3673 from chenghao-intel/spark_4825 and squashes the following
commits:
e8cbd56 [Cheng Hao] alternate the pattern matching order for logical
plan:CTAS
e004895 [Cheng Hao] fix bug
commit da13b63936b26d11dd8c17afd133db9405731c21
Author: Sasaki Toru <[email protected]>
Date: 2014-12-12T06:54:21Z
[SPARK-4742][SQL] The name of Parquet File generated by
AppendingParquetOutputFormat should be zero padded
When I use Parquet File as a output file using
ParquetOutputFormat#getDefaultWorkFile, the file name is not zero padded while
RDD#saveAsText does zero padding.
Author: Sasaki Toru <[email protected]>
Closes #3602 from sasakitoa/parquet-zeroPadding and squashes the following
commits:
6b0e58f [Sasaki Toru] Merge branch 'master' of
git://github.com/apache/spark into parquet-zeroPadding
20dc79d [Sasaki Toru] Fixed the name of Parquet File generated by
AppendingParquetOutputFormat
commit 53b74f57fc22fc5645f1e313349542cf77044b13
Author: Daoyuan Wang <[email protected]>
Date: 2014-12-12T06:56:42Z
[SPARK-4829] [SQL] add rule to fold count(expr) if expr is not null
Author: Daoyuan Wang <[email protected]>
Closes #3676 from adrian-wang/countexpr and squashes the following commits:
dc5765b [Daoyuan Wang] add rule to fold count(expr) if expr is not null
commit 7cdebe97d433fc4346e6b983ae2dd408608b7ccc
Author: Peter Klipfel <[email protected]>
Date: 2014-12-14T08:01:16Z
fixed spelling errors in documentation
changed "form" to "from" in 3 documentation entries for Kafka integration
Author: Peter Klipfel <[email protected]>
Closes #3691 from peterklipfel/master and squashes the following commits:
0fe7fc5 [Peter Klipfel] fixed spelling errors in documentation
commit d89b26c97941bc9c026671a1380ca18036e26e3f
Author: Patrick Wendell <[email protected]>
Date: 2014-12-15T18:54:45Z
HOTFIX: Disabling failing block manager test
commit 7cf4d7a4dedb5fe3bd6e108990f70bd848e68066
Author: Yuu ISHIKAWA <[email protected]>
Date: 2014-12-15T21:44:15Z
[SPARK-4494][mllib] IDFModel.transform() add support for single vector
I improved `IDFModel.transform` to allow using a single vector.
[[SPARK-4494] IDFModel.transform() add support for single vector - ASF
JIRA](https://issues.apache.org/jira/browse/SPARK-4494)
Author: Yuu ISHIKAWA <[email protected]>
Closes #3603 from yu-iskw/idf and squashes the following commits:
256ff3d [Yuu ISHIKAWA] Fix typo
a3bf566 [Yuu ISHIKAWA] - Fix typo - Optimize import order - Aggregate the
assertion tests - Modify `IDFModel.transform` API for pyspark
d25e49b [Yuu ISHIKAWA] Add the implementation of `IDFModel.transform` for a
term frequency vector
commit 8347fbe06ca08b83c1e4a4291e326b394f5eb24c
Author: Josh Rosen <[email protected]>
Date: 2014-12-15T22:33:43Z
[SPARK-4826] Fix generation of temp file names in WAL tests
This PR should fix SPARK-4826, an issue where a bug in how we generate
temp. file names was causing spurious test failures in the write ahead log
suites.
Closes #3695.
Closes #3701.
Author: Josh Rosen <[email protected]>
Closes #3704 from JoshRosen/SPARK-4826 and squashes the following commits:
f2307f5 [Josh Rosen] Use Spark Utils class for directory creation/deletion
a693ddb [Josh Rosen] remove unused Random import
b275e41 [Josh Rosen] Move creation of temp. dir to beforeEach/afterEach.
9362919 [Josh Rosen] [SPARK-4826] Fix bug in generation of temp file names.
in WAL suites.
86c1944 [Josh Rosen] Revert "HOTFIX: Disabling failing block manager test"
commit 9d3276737facf5838eec8e08404a2fd7f52940db
Author: Ilya Ganelin <[email protected]>
Date: 2014-12-15T22:51:15Z
[SPARK-1037] The name of findTaskFromList & findTask in
TaskSetManager.scala is confusing
Hi all - I've renamed the methods referenced in this JIRA to clarify that
they modify the provided arrays (find vs. deque).
Author: Ilya Ganelin <[email protected]>
Closes #3665 from ilganeli/SPARK-1037B and squashes the following commits:
64c177c [Ilya Ganelin] Renamed deque to dequeue
f27d85e [Ilya Ganelin] Renamed private methods to clarify that they modify
the provided parameters
683482a [Ilya Ganelin] Renamed private methods to clarify that they modify
the provided parameters
commit be7d6ea34e5f38c5f77d00014a6425224d505e4e
Author: Ryan Williams <[email protected]>
Date: 2014-12-15T22:52:17Z
[SPARK-4668] Fix some documentation typos.
Author: Ryan Williams <[email protected]>
Closes #3523 from ryan-williams/tweaks and squashes the following commits:
d2eddaa [Ryan Williams] code review feedback
ce27fc1 [Ryan Williams] CoGroupedRDD comment nit
c6cfad9 [Ryan Williams] remove unnecessary if statement
b74ea35 [Ryan Williams] comment fix
b0221f0 [Ryan Williams] fix a gendered pronoun
c71ffed [Ryan Williams] use names on a few boolean parameters
89954aa [Ryan Williams] clarify some comments in {Security,Shuffle}Manager
e465dac [Ryan Williams] Saved building-spark.md with Dillinger.io
83e8358 [Ryan Williams] fix pom.xml typo
dc4662b [Ryan Williams] typo fixes in tuning.md, configuration.md
commit 6c185e926d26bd6ca25cab85e431641cc2e9a4a8
Author: Sean Owen <[email protected]>
Date: 2014-12-16T00:06:15Z
SPARK-785 [CORE] ClosureCleaner not invoked on most PairRDDFunctions
This looked like perhaps a simple and important one. `combineByKey` looks
like it should clean its arguments' closures, and that in turn covers
apparently all remaining functions in `PairRDDFunctions` which delegate to it.
Author: Sean Owen <[email protected]>
Closes #3690 from srowen/SPARK-785 and squashes the following commits:
8df68fe [Sean Owen] Clean context of most remaining functions in
PairRDDFunctions, which ultimately call combineByKey
commit cb465e0fbb3da3b8f4abaf49d071eb7a9692a6cc
Author: wangfei <[email protected]>
Date: 2014-12-16T00:46:21Z
[Minor][Core] fix comments in MapOutputTracker
Using driver and executor in the comments of ```MapOutputTracker``` is more
clear.
Author: wangfei <[email protected]>
Closes #3700 from scwf/commentFix and squashes the following commits:
aa68524 [wangfei] master and worker should be driver and executor
commit e147c42320c3a8b89940ec93110e6c83fc9954ba
Author: Sean Owen <[email protected]>
Date: 2014-12-16T01:12:05Z
SPARK-4814 [CORE] Enable assertions in SBT, Maven tests / AssertionError
from Hive's LazyBinaryInteger
This enables assertions for the Maven and SBT build, but overrides the Hive
module to not enable assertions.
Author: Sean Owen <[email protected]>
Closes #3692 from srowen/SPARK-4814 and squashes the following commits:
caca704 [Sean Owen] Disable assertions just for Hive
f71e783 [Sean Owen] Enable assertions for SBT and Maven build
commit f2c7c90830ecca9603ce53c43ec26dc0a3cdbad2
Author: meiyoula <[email protected]>
Date: 2014-12-16T06:30:18Z
[SPARK-4792] Add error message when making local dir unsuccessfully
Author: meiyoula <[email protected]>
Closes #3635 from XuTingjun/master and squashes the following commits:
dd1c66d [meiyoula] when old is deleted, it will throw an exception where
call it
2a55bc2 [meiyoula] Update DiskBlockManager.scala
1483a4a [meiyoula] Delete multiple retries to make dir
67f7902 [meiyoula] Try some times to make dir maybe more reasonable
1c51a0c [meiyoula] Update DiskBlockManager.scala
commit b1f3e00a550c65ed11dbdccc92ef0ebecebb225a
Author: Davies Liu <[email protected]>
Date: 2014-12-16T06:58:26Z
[SPARK-4841] fix zip with textFile()
UTF8Deserializer can not be used in BatchedSerializer, so always use
PickleSerializer() when change batchSize in zip().
Also, if two RDD have the same batch size already, they did not need
re-serialize any more.
Author: Davies Liu <[email protected]>
Closes #3706 from davies/fix_4841 and squashes the following commits:
20ce3a3 [Davies Liu] fix bug in _reserialize()
e3ebf7c [Davies Liu] add comment
379d2c8 [Davies Liu] fix zip with textFile()
commit 6ce251a008345a26ab94fc6ee567e070523560b9
Author: Davies Liu <[email protected]>
Date: 2014-12-16T19:19:36Z
[SPARK-4437] update doc for WholeCombineFileRecordReader
update doc for WholeCombineFileRecordReader
Author: Davies Liu <[email protected]>
Author: Josh Rosen <[email protected]>
Closes #3301 from davies/fix_doc and squashes the following commits:
1d7422f [Davies Liu] Merge pull request #2 from
JoshRosen/whole-text-file-cleanup
dc3d21a [Josh Rosen] More genericization in
ConfigurableCombineFileRecordReader.
95d13eb [Davies Liu] address comment
bf800b9 [Davies Liu] update doc for WholeCombineFileRecordReader
commit 01d123ffbf3e5e517ac1ad8f2207cfebac2e8bdd
Author: jbencook <[email protected]>
Date: 2014-12-16T19:37:23Z
[SPARK-4855][mllib] testing the Chi-squared hypothesis test
This PR tests the pyspark Chi-squared hypothesis test from this commit:
c8abddc5164d8cf11cdede6ab3d5d1ea08028708 and moves some of the error messaging
in to python.
It is a port of the Scala tests here:
[HypothesisTestSuite.scala](https://github.com/apache/spark/blob/master/mllib/src/test/scala/org/apache/spark/mllib/stat/HypothesisTestSuite.scala)
Hopefully, SPARK-2980 can be closed.
Author: jbencook <[email protected]>
Closes #3679 from jbencook/master and squashes the following commits:
44078e0 [jbencook] checking that bad input throws the correct exceptions
f12ee10 [jbencook] removing checks for ValueError since input tests are on
the Scala side
7536cf1 [jbencook] removing python checks for invalid input
a17ee84 [jbencook] [SPARK-2980][mllib] adding unit tests for the pyspark
chi-squared test
3aeb0d9 [jbencook] [SPARK-2980][mllib] bringing Chi-squared error messages
to the python side
commit d52330cdfffb689a8bc80e277c5e30556459cc0a
Author: Mike Jennings <[email protected]>
Date: 2014-12-16T20:13:21Z
[SPARK-3405] add subnet-id and vpc-id options to spark_ec2.py
Based on this gist:
https://gist.github.com/amar-analytx/0b62543621e1f246c0a2
We use security group ids instead of security group to get around this
issue:
https://github.com/boto/boto/issues/350
Author: Mike Jennings <[email protected]>
Author: Mike Jennings <[email protected]>
Closes #2872 from mvj101/SPARK-3405 and squashes the following commits:
be9cb43 [Mike Jennings] `pep8 spark_ec2.py` runs cleanly.
4dc6756 [Mike Jennings] Remove duplicate comment
731d94c [Mike Jennings] Update for code review.
ad90a36 [Mike Jennings] Merge branch 'master' of
https://github.com/apache/spark into SPARK-3405
1ebffa1 [Mike Jennings] Merge branch 'master' into SPARK-3405
52aaeec [Mike Jennings] [SPARK-3405] add subnet-id and vpc-id options to
spark_ec2.py
commit 47a43c0475676ab1a0a8abd800b00dbc45e1c49d
Author: Judy Nash <[email protected]>
Date: 2014-12-16T20:37:26Z
[SQL] SPARK-4700: Add HTTP protocol spark thrift server
Add HTTP protocol support and test cases to spark thrift server, so users
can deploy thrift server in both TCP and http mode.
Author: Judy Nash <[email protected]>
Author: judynash <[email protected]>
Closes #3672 from judynash/master and squashes the following commits:
526315d [Judy Nash] correct spacing on startThriftServer method
31a6520 [Judy Nash] fix code style issues and update sql programming guide
format issue
47bf87e [Judy Nash] modify withJdbcStatement method definition to meet less
than 100 line length
2e9c11c [Judy Nash] add thrift server in http mode documentation on sql
programming guide
1cbd305 [Judy Nash] Merge remote-tracking branch 'upstream/master'
2b1d312 [Judy Nash] updated http thrift server support based on feedback
377532c [judynash] add HTTP protocol spark thrift server
commit 4b5fd5f1c632843434b6e61ea263471e75998958
Author: Peter Vandenabeele <[email protected]>
Date: 2014-12-16T21:57:55Z
[DOCS][SQL] Add a Note on jsonFile having separate JSON objects per line
* This commit hopes to avoid the confusion I faced when trying
to submit a regular, valid multi-line JSON file, also see
http://apache-spark-user-list.1001560.n3.nabble.com/Loading-JSON-Dataset-fails-with-com-fasterxml-jackson-databind-JsonMappingException-td20041.html
Author: Peter Vandenabeele <[email protected]>
Closes #3517 from petervandenabeele/pv-docs-note-on-jsonFile-format/01 and
squashes the following commits:
1f98e52 [Peter Vandenabeele] Revert to people.json and simple Note text
6b6e062 [Peter Vandenabeele] Change the "JSON" connotation to "txt"
fca7dfb [Peter Vandenabeele] Add a Note on jsonFile having separate JSON
objects per line
commit 88bef9990fb3ef3cd53faf780c618a24ffd0a42d
Author: jerryshao <[email protected]>
Date: 2014-12-16T22:08:28Z
[SPARK-4847][SQL]Fix "extraStrategies cannot take effect in SQLContext"
issue
Author: jerryshao <[email protected]>
Closes #3698 from jerryshao/SPARK-4847 and squashes the following commits:
4741130 [jerryshao] Make later added extraStrategies effect when calling
strategies
commit 0574fd8c2012da8fccb95f4f519426deb807d906
Author: zsxwing <[email protected]>
Date: 2014-12-16T22:13:40Z
[SPARK-4812][SQL] Fix the initialization issue of 'codegenEnabled'
The problem is `codegenEnabled` is `val`, but it uses a `val` `sqlContext`,
which can be override by subclasses. Here is a simple example to show this
issue.
```Scala
scala> :paste
// Entering paste mode (ctrl-D to finish)
abstract class Foo {
protected val sqlContext = "Foo"
val codegenEnabled: Boolean = {
println(sqlContext) // it will call subclass's `sqlContext` which has
not yet been initialized.
if (sqlContext != null) {
true
} else {
false
}
}
}
class Bar extends Foo {
override val sqlContext = "Bar"
}
println(new Bar().codegenEnabled)
// Exiting paste mode, now interpreting.
null
false
defined class Foo
defined class Bar
```
We should make `sqlContext` `final` to prevent subclasses from overriding
it incorrectly.
Author: zsxwing <[email protected]>
Closes #3660 from zsxwing/SPARK-4812 and squashes the following commits:
1cbb623 [zsxwing] Make `sqlContext` final to prevent subclasses from
overriding it incorrectly
commit d136891dc408795aacdcd11634e2e968c28bc96b
Author: Holden Karau <[email protected]>
Date: 2014-12-16T22:37:04Z
SPARK-4767: Add support for launching in a specified placement group to
spark_ec2
Placement groups are cool and all the cool kids are using them. Lets add
support for them to spark_ec2.py because I'm lazy
Author: Holden Karau <[email protected]>
Closes #3623 from
holdenk/SPARK-4767-add-support-for-launching-in-a-specified-placement-group-to-spark-ec2-scripts
and squashes the following commits:
111a5fd [Holden Karau] merge in master
70ace25 [Holden Karau] Placement groups are cool and all the cool kids are
using them. Lets add support for them to spark_ec2.py because I'm lazy
commit 010e1a66fd2a0843704f52232e1d5dbcc6e779a1
Author: wangxiaojing <[email protected]>
Date: 2014-12-16T22:45:56Z
[SPARK-4527][SQl]Add BroadcastNestedLoopJoin operator selection testsuite
In `JoinSuite` add BroadcastNestedLoopJoin operator selection testsuite
Author: wangxiaojing <[email protected]>
Closes #3395 from wangxiaojing/SPARK-4527 and squashes the following
commits:
ea0e495 [wangxiaojing] change style
53c3952 [wangxiaojing] Add BroadcastNestedLoopJoin operator selection
testsuite
commit 6d0d264b6b2b2d080a74b855a01e47ad94150569
Author: tianyi <[email protected]>
Date: 2014-12-16T23:22:29Z
[SPARK-4483][SQL]Optimization about reduce memory costs during the
HashOuterJoin
In `HashOuterJoin.scala`, spark read data from both side of join operation
before zip them together. It is a waste for memory. We are trying to read data
from only one side, put them into a hashmap, and then generate the `JoinedRow`
with data from other side one by one.
Currently, we could only do this optimization for `left outer join` and
`right outer join`. For `full outer join`, we will do something in another
issue.
for
table test_csv contains 1 million records
table dim_csv contains 10 thousand records
SQL:
`select * from test_csv a left outer join dim_csv b on a.key = b.key`
the result is:
master:
```
CSV: 12671 ms
CSV: 9021 ms
CSV: 9200 ms
Current Mem Usage:787788984
```
after patchï¼
```
CSV: 10382 ms
CSV: 7543 ms
CSV: 7469 ms
Current Mem Usage:208145728
```
Author: tianyi <[email protected]>
Author: tianyi <[email protected]>
Closes #3375 from tianyi/SPARK-4483 and squashes the following commits:
72a8aec [tianyi] avoid having mutable state stored inside of the task
99c5c97 [tianyi] performance optimization
d2f94d7 [tianyi] fix bug: missing output when the join-key is null.
2be45d1 [tianyi] fix spell bug
1f2c6f1 [tianyi] remove commented codes
a676de6 [tianyi] optimize some codes
9e7d5b5 [tianyi] remove commented old codes
838707d [tianyi] Optimization about reduce memory costs during the
HashOuterJoin
commit b4abf0f384513b8403df29adf8f0b9a91a79bb7c
Author: Michael Armbrust <[email protected]>
Date: 2014-12-16T23:31:19Z
[SPARK-4827][SQL] Fix resolution of deeply nested Project(attr,
Project(Star,...)).
Since `AttributeReference` resolution and `*` expansion are currently in
separate rules, each pair requires a full iteration instead of being able to
resolve in a single pass. Since its pretty easy to construct queries that have
many of these in a row, I combine them into a single rule in this PR.
Author: Michael Armbrust <[email protected]>
Closes #3674 from marmbrus/projectStars and squashes the following commits:
d83d6a1 [Michael Armbrust] Fix resolution of deeply nested Project(attr,
Project(Star,...)).
commit b2ff56dd71e11a4f0fac1dda1c7492e5eeeaafd0
Author: Jacky Li <[email protected]>
Date: 2014-12-16T23:34:59Z
[SPARK-4269][SQL] make wait time configurable in BroadcastHashJoin
In BroadcastHashJoin, currently it is using a hard coded value (5 minutes)
to wait for the execution and broadcast of the small table.
In my opinion, it should be a configurable value since broadcast may exceed
5 minutes in some case, like in a busy/congested network environment.
Author: Jacky Li <[email protected]>
Closes #3133 from jackylk/timeout-config and squashes the following commits:
733ac08 [Jacky Li] add spark.sql.broadcastTimeout in SQLConf.scala
557acd4 [Jacky Li] switch to sqlContext.getConf
81a5e20 [Jacky Li] make wait time configurable in BroadcastHashJoin
commit b58814c59bd172a817d42b283b107a87eafde15b
Author: Andrew Or <[email protected]>
Date: 2014-12-17T01:55:27Z
[Release] Major improvements to generate contributors script
This commit introduces several major improvements to the script
that generates the contributors list for release notes, notably:
(1) Use release tags instead of a range of commits. Across branches,
commits are not actually strictly two-dimensional, and so it is not
sufficient to specify a start hash and an end hash. Otherwise, we
end up counting commits that were already merged in an older branch.
(2) Match PR numbers in addition to commit hashes. This is related
to the first point in that if a PR is already merged in an older
minor release tag, it should be filtered out here. This requires us
to do some intelligent regex parsing on the commit description in
addition to just relying on the GitHub API.
(3) Relax author validity check. The old code fails on a name that
has many middle names, for instance. The test was just too strict.
(4) Use GitHub authentication. This allows us to make far more
requests through the GitHub API than before (5000 as opposed to 60
per hour).
(5) Translate from Github username, not commit author name. This is
important because the commit author name is not always configured
correctly by the user. For instance, the username "falaki" used to
resolve to just "Hossein", which was treated as a github username
and translated to something else that is completely arbitrary.
(6) Add an option to use the untranslated name. If there is not
a satisfactory candidate to replace the untranslated name with,
at least allow the user to not translate it.
commit 5ef3a1b37b6ee385f84431fd2c17ace1b119a67d
Author: Andrew Or <[email protected]>
Date: 2014-12-17T03:28:43Z
[Release] Cache known author translations locally
This bypasses unnecessary calls to the Github and JIRA API.
Additionally, having a local cache allows us to remember names
that we had to manually discover ourselves.
commit bdf561cc4ce2c4f5b398b38b57cf33cc0388b156
Author: Cheng Lian <[email protected]>
Date: 2014-12-17T05:16:03Z
[SPARK-4798][SQL] A new set of Parquet testing API and test suites
This PR provides a set Parquet testing API (see trait `ParquetTest`) that
enables developers to write more concise test cases. A new set of Parquet test
suites built upon this API are added and aim to replace the old
`ParquetQuerySuite`. To avoid potential merge conflicts, old testing code are
not removed yet. The following classes can be safely removed after most Parquet
related PRs are handled:
- `ParquetQuerySuite`
- `ParquetTestData`
<!-- Reviewable:start -->
[<img src="https://reviewable.io/review_button.png" height=40 alt="Review
on Reviewable"/>](https://reviewable.io/reviews/apache/spark/3644)
<!-- Reviewable:end -->
Author: Cheng Lian <[email protected]>
Closes #3644 from liancheng/parquet-tests and squashes the following
commits:
800e745 [Cheng Lian] Enforces ordering of test output
3bb8731 [Cheng Lian] Refactors HiveParquetSuite
aa2cb2e [Cheng Lian] Decouples ParquetTest and TestSQLContext
7b43a68 [Cheng Lian] Updates ParquetTest Scaladoc
7f07af0 [Cheng Lian] Adds a new set of Parquet test suites
commit 5db941c9d05c575c5414b544e3ee6fb6491cec38
Author: Cheng Hao <[email protected]>
Date: 2014-12-17T05:18:39Z
[SPARK-4744] [SQL] Short circuit evaluation for AND & OR in CodeGen
Author: Cheng Hao <[email protected]>
Closes #3606 from chenghao-intel/codegen_short_circuit and squashes the
following commits:
f466303 [Cheng Hao] short circuit for AND & OR
----
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]