I think you meant to use the "--files" to deploy the DLLs. I gave a try, but it did not work.
>From the Spark UI, Environment tab, I can see spark.yarn.dist.files file:/c:/openblas/libgcc_s_seh-1.dll,file:/c:/openblas/libblas3.dll,file:/c:/openblas/libgfortran-3.dll,file:/c:/openblas/liblapack3.dll,file:/c:/openblas/libquadmath-0.dll I think my DLLs are all deployed. But I still got the warn message that native BLAS library cannot be load. And idea? Thanks, David On Wed, Mar 25, 2015 at 5:40 AM DB Tsai <dbt...@dbtsai.com> wrote: > I would recommend to upload those jars to HDFS, and use add jars > option in spark-submit with URI from HDFS instead of URI from local > filesystem. Thus, it can avoid the problem of fetching jars from > driver which can be a bottleneck. > > Sincerely, > > DB Tsai > ------------------------------------------------------- > Blog: https://www.dbtsai.com > > > On Tue, Mar 24, 2015 at 4:13 AM, Xi Shen <davidshe...@gmail.com> wrote: > > Hi, > > > > I am doing ML using Spark mllib. However, I do not have full control to > the > > cluster. I am using Microsoft Azure HDInsight > > > > I want to deploy the BLAS or whatever required dependencies to accelerate > > the computation. But I don't know how to deploy those DLLs when I submit > my > > JAR to the cluster. > > > > I know how to pack those DLLs into a jar. The real challenge is how to > let > > the system find them... > > > > > > Thanks, > > David > > >