Github user mengxr commented on a diff in the pull request:

    https://github.com/apache/spark/pull/21492#discussion_r192892747
  
    --- Diff: 
mllib/src/test/scala/org/apache/spark/ml/clustering/LDASuite.scala ---
    @@ -36,9 +36,9 @@ object LDASuite {
           vocabSize: Int): DataFrame = {
         val avgWC = 1  // average instances of each word in a doc
         val sc = spark.sparkContext
    -    val rng = new java.util.Random()
    -    rng.setSeed(1)
         val rdd = sc.parallelize(1 to rows).map { i =>
    +      val rng = new java.util.Random()
    --- End diff --
    
    minor: you can use `new java.util.random(i)` direct


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to