Github user mengxr commented on a diff in the pull request:
https://github.com/apache/spark/pull/21492#discussion_r192892747
--- Diff:
mllib/src/test/scala/org/apache/spark/ml/clustering/LDASuite.scala ---
@@ -36,9 +36,9 @@ object LDASuite {
vocabSize: Int): DataFrame = {
val avgWC = 1 // average instances of each word in a doc
val sc = spark.sparkContext
- val rng = new java.util.Random()
- rng.setSeed(1)
val rdd = sc.parallelize(1 to rows).map { i =>
+ val rng = new java.util.Random()
--- End diff --
minor: you can use `new java.util.random(i)` direct
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]