[ 
https://issues.apache.org/jira/browse/MAHOUT-1653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14616207#comment-14616207
 ] 

ASF GitHub Bot commented on MAHOUT-1653:
----------------------------------------

Github user dlyubimov commented on a diff in the pull request:

    https://github.com/apache/mahout/pull/146#discussion_r34008915
  
    --- Diff: 
spark-shell/src/main/scala/org/apache/mahout/sparkbindings/shell/MahoutSparkILoop.scala
 ---
    @@ -48,55 +77,81 @@ class MahoutSparkILoop extends SparkILoop {
     
         conf.set("spark.executor.memory", "1g")
     
    -    sparkContext = mahoutSparkContext(
    +    _interp.sparkContext = mahoutSparkContext(
           masterUrl = master,
           appName = "Mahout Spark Shell",
           customJars = jars,
           sparkConf = conf
         )
     
    --- End diff --
    
    ok what i meant is that this should really be saved as MahoutDistributed 
context :
    
        private val sdc = mahoutSparkContext(...) 
    
    which is already of type SparkDistributedContext (mahout type), right?
    
    and then use to initialize `sdc = this.sdc` and `sc = this.sdc.sc` later... 
We still seem to be crating DistributedSparkContext twice here, which is 
probably not a big deal but feels a bit hacky...


> Spark 1.3
> ---------
>
>                 Key: MAHOUT-1653
>                 URL: https://issues.apache.org/jira/browse/MAHOUT-1653
>             Project: Mahout
>          Issue Type: Dependency upgrade
>            Reporter: Andrew Musselman
>            Assignee: Andrew Palumbo
>             Fix For: 0.11.0
>
>
> Support Spark 1.3



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to