Hey Nikaash,
thanks for the help. command is running but its not aasigning any worker to
job. IT keep keeps showwing this following message.
Initial job has not accepted any resources; check your cluster UI to ensure
that workers are registered and have sufficient resources

On Thu, Apr 14, 2016 at 3:48 PM, Nikaash Puri <nikaashp...@gmail.com> wrote:

> my guess is that its not initialising the Configuration correctly since
> the spark master is by default local. My guess is that to read from HDFS
> you’d have to set up a Spark cluster and specify the master URL using the
> —master flag.
>
> Else, set up a Spark on YARN deployment using CDH. Then specify the master
> as —yarn-cluster. That should also work.
> > On 14-Apr-2016, at 3:43 PM, Rohit Jain <rohitkjai...@gmail.com> wrote:
> >
> > Hello Nikaash
> >
> > I did this.
> >
> > /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/bin/mahout
> > spark-itemsimilarity -i hdfs://localhost:9000/spdemo.txt -o
> > hdfs://localhost:9000/spout/ --filter1 like -fc 1 -ic 2
> > -D:spark.io.compression.codec=lzf
> >
> > but it throws this exception
> >
> > Exception in thread "main" java.lang.IllegalArgumentException: Wrong FS:
> > hdfs://localhost:9000/spdemo1.txt, expected: file:///
> >
> > On Thu, Apr 14, 2016 at 3:41 PM, Nikaash Puri <nikaashp...@gmail.com>
> wrote:
> >
> >> Hi,
> >>
> >> You’d need to specify the complete path along with HDFS URI. for eg.
> >> hdfs://namenode:<port>/path/to/file.
> >>
> >> - Nikaash
> >>> On 14-Apr-2016, at 3:39 PM, Rohit Jain <rohitkjai...@gmail.com> wrote:
> >>>
> >>> Hey Nikaash,
> >>>
> >>> I ran this command but its working local files not for hadoop input
> file.
> >>> How to make it work on hadoop uri?
> >>>
> >>> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/bin/mahout
> >>> spark-itemsimilarity -i spdemo.txt -o /spout --filter1 like -fc 1 -ic 2
> >>> -D:spark.io.compression.codec=lzf
> >>>
> >>> On Thu, Apr 14, 2016 at 3:19 PM, Nikaash Puri <nikaashp...@gmail.com>
> >> wrote:
> >>>
> >>>> Cool. Even if you have problems setting up the library, just change
> the
> >>>> compression codec from snappy to lzf
> (-D:spark.io.compression.codec=lzf)
> >>>> like in the command. It might hurt performance, but it’ll get it to
> work
> >>>>
> >>>> - Nikaash
> >>>>> On 14-Apr-2016, at 3:17 PM, Rohit Jain <rohitkjai...@gmail.com>
> wrote:
> >>>>>
> >>>>> yes. i am downloading the that library and putting it java/lib/ path.
> >>>>> Anythingelse need to be done?
> >>>>>
> >>>>> On Thu, Apr 14, 2016 at 3:08 PM, Nikaash Puri <nikaashp...@gmail.com
> >
> >>>> wrote:
> >>>>>
> >>>>>> Hi Rohit,
> >>>>>>
> >>>>>> It’s a snappy library issue. Are you running on Mac?
> >>>>>> Anyhow, run it like this:
> >>>>>> ./bin/mahout spark-itemsimilarity --input <in> --output <out>
> >>>>>> -D:spark.io.compression.codec=lzf
> >>>>>>
> >>>>>> Nikaash
> >>>>>>
> >>>>>>> On 14-Apr-2016, at 12:25 PM, Rohit Jain <rohitkjai...@gmail.com>
> >>>> wrote:
> >>>>>>>
> >>>>>>> Hello everyone,
> >>>>>>> I am new to mahout. I am trying to run this command
> >>>>>>>
> >>>>>>> /Documents/apache-mahout-distribution-0.11.2/bin/mahout
> >>>>>>> spark-itemsimilarity -i spdemo1.txt -o /spout --filter1 like -fc 1
> >> -ic
> >>>> 2
> >>>>>>>
> >>>>>>> before running the command i uploaded file to hdfs
> >>>>>>>
> >>>>>>> bin/hadoop dfs -put spdemo.txt /spdemo1.txt
> >>>>>>>
> >>>>>>> but i am geting this long list of error.Do I need to install spark
> >> too?
> >>>>>>>
> >>>>>>>
> >>>>>>> SLF4J: Class path contains multiple SLF4J bindings.
> >>>>>>>
> >>>>>>> SLF4J: Found binding in
> >>>>>>>
> >>>>>>
> >>>>
> >>
> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-examples-0.11.2-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >>>>>>>
> >>>>>>> SLF4J: Found binding in
> >>>>>>>
> >>>>>>
> >>>>
> >>
> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-mr-0.11.2-job.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >>>>>>>
> >>>>>>> SLF4J: Found binding in
> >>>>>>>
> >>>>>>
> >>>>
> >>
> [jar:file:/Users/rohitjain/Documents/spark-1.6.1-bin-hadoop2.6/lib/spark-assembly-1.6.1-hadoop2.6.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >>>>>>>
> >>>>>>> SLF4J: Found binding in
> >>>>>>>
> >>>>>>
> >>>>
> >>
> [jar:file:/Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/lib/slf4j-log4j12-1.7.12.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> >>>>>>>
> >>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for
> an
> >>>>>>> explanation.
> >>>>>>>
> >>>>>>> SLF4J: Actual binding is of type
> [org.slf4j.impl.Log4jLoggerFactory]
> >>>>>>>
> >>>>>>> 16/04/14 10:53:09 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> may be removed in the future. Please use
> spark.kryoserializer.buffer
> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
> >>>>>>> previously specified as '0.064'. Fractional values are no longer
> >>>>>> accepted.
> >>>>>>> To specify the equivalent now, one may use '64k'.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:09 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> and may be removed in the future. Please use the new key
> >>>>>>> 'spark.kryoserializer.buffer' instead.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:10 INFO SparkContext: Running Spark version 1.6.1
> >>>>>>>
> >>>>>>> 16/04/14 10:53:10 WARN NativeCodeLoader: Unable to load
> native-hadoop
> >>>>>>> library for your platform... using builtin-java classes where
> >>>> applicable
> >>>>>>>
> >>>>>>> 16/04/14 10:53:10 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> may be removed in the future. Please use
> spark.kryoserializer.buffer
> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
> >>>>>>> previously specified as '0.064'. Fractional values are no longer
> >>>>>> accepted.
> >>>>>>> To specify the equivalent now, one may use '64k'.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:10 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> and may be removed in the future. Please use the new key
> >>>>>>> 'spark.kryoserializer.buffer' instead.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:10 INFO SecurityManager: Changing view acls to:
> >>>> rohitjain
> >>>>>>>
> >>>>>>> 16/04/14 10:53:10 INFO SecurityManager: Changing modify acls to:
> >>>>>> rohitjain
> >>>>>>>
> >>>>>>> 16/04/14 10:53:10 INFO SecurityManager: SecurityManager:
> >> authentication
> >>>>>>> disabled; ui acls disabled; users with view permissions:
> >>>> Set(rohitjain);
> >>>>>>> users with modify permissions: Set(rohitjain)
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> may be removed in the future. Please use
> spark.kryoserializer.buffer
> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
> >>>>>>> previously specified as '0.064'. Fractional values are no longer
> >>>>>> accepted.
> >>>>>>> To specify the equivalent now, one may use '64k'.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> and may be removed in the future. Please use the new key
> >>>>>>> 'spark.kryoserializer.buffer' instead.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> may be removed in the future. Please use
> spark.kryoserializer.buffer
> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
> >>>>>>> previously specified as '0.064'. Fractional values are no longer
> >>>>>> accepted.
> >>>>>>> To specify the equivalent now, one may use '64k'.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> and may be removed in the future. Please use the new key
> >>>>>>> 'spark.kryoserializer.buffer' instead.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO Utils: Successfully started service
> >>>> 'sparkDriver'
> >>>>>> on
> >>>>>>> port 64062.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO Slf4jLogger: Slf4jLogger started
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO Remoting: Starting remoting
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO Remoting: Remoting started; listening on
> >>>> addresses
> >>>>>>> :[akka.tcp://sparkDriverActorSystem@192.168.1.122:64063]
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO Utils: Successfully started service
> >>>>>>> 'sparkDriverActorSystem' on port 64063.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO SparkEnv: Registering MapOutputTracker
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> may be removed in the future. Please use
> spark.kryoserializer.buffer
> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
> >>>>>>> previously specified as '0.064'. Fractional values are no longer
> >>>>>> accepted.
> >>>>>>> To specify the equivalent now, one may use '64k'.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> and may be removed in the future. Please use the new key
> >>>>>>> 'spark.kryoserializer.buffer' instead.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> may be removed in the future. Please use
> spark.kryoserializer.buffer
> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
> >>>>>>> previously specified as '0.064'. Fractional values are no longer
> >>>>>> accepted.
> >>>>>>> To specify the equivalent now, one may use '64k'.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> and may be removed in the future. Please use the new key
> >>>>>>> 'spark.kryoserializer.buffer' instead.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO SparkEnv: Registering BlockManagerMaster
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO DiskBlockManager: Created local directory at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/blockmgr-6cdfd6f1-56c8-445c-95e0-70305b0f5f5a
> >>>>>>>
> >>>>>>> 16/04/14 10:53:11 INFO MemoryStore: MemoryStore started with
> capacity
> >>>>>> 2.4 GB
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO SparkEnv: Registering
> OutputCommitCoordinator
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO Server: jetty-8.y.z-SNAPSHOT
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO AbstractConnector: Started
> >>>>>>> SelectChannelConnector@0.0.0.0:4040
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service
> 'SparkUI'
> >> on
> >>>>>>> port 4040.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO SparkUI: Started SparkUI at
> >>>>>> http://192.168.1.122:4040
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO HttpFileServer: HTTP File server directory
> is
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b/httpd-11f32419-5f8a-4b61-9654-9ea2d32cab5d
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO HttpServer: Starting HTTP Server
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO Server: jetty-8.y.z-SNAPSHOT
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO AbstractConnector: Started
> >>>>>>> SocketConnector@0.0.0.0:64064
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service 'HTTP
> file
> >>>>>>> server' on port 64064.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-hdfs-0.11.2.jar
> >>>>>>> at http://192.168.1.122:64064/jars/mahout-hdfs-0.11.2.jar with
> >>>> timestamp
> >>>>>>> 1460611392248
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-math-0.11.2.jar
> >>>>>>> at http://192.168.1.122:64064/jars/mahout-math-0.11.2.jar with
> >>>> timestamp
> >>>>>>> 1460611392253
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-math-scala_2.10-0.11.2.jar
> >>>>>>> at
> http://192.168.1.122:64064/jars/mahout-math-scala_2.10-0.11.2.jar
> >>>>>> with
> >>>>>>> timestamp 1460611392255
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-spark_2.10-0.11.2-dependency-reduced.jar
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> http://192.168.1.122:64064/jars/mahout-spark_2.10-0.11.2-dependency-reduced.jar
> >>>>>>> with timestamp 1460611392303
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO SparkContext: Added JAR
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /Users/rohitjain/Documents/apache-mahout-distribution-0.11.2/mahout-spark_2.10-0.11.2.jar
> >>>>>>> at http://192.168.1.122:64064/jars/mahout-spark_2.10-0.11.2.jar
> with
> >>>>>>> timestamp 1460611392304
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> may be removed in the future. Please use
> spark.kryoserializer.buffer
> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
> >>>>>>> previously specified as '0.064'. Fractional values are no longer
> >>>>>> accepted.
> >>>>>>> To specify the equivalent now, one may use '64k'.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> and may be removed in the future. Please use the new key
> >>>>>>> 'spark.kryoserializer.buffer' instead.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> may be removed in the future. Please use
> spark.kryoserializer.buffer
> >>>>>>> instead. The default value for spark.kryoserializer.buffer.mb was
> >>>>>>> previously specified as '0.064'. Fractional values are no longer
> >>>>>> accepted.
> >>>>>>> To specify the equivalent now, one may use '64k'.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 WARN SparkConf: The configuration key
> >>>>>>> 'spark.kryoserializer.buffer.mb' has been deprecated as of Spark
> 1.4
> >>>> and
> >>>>>>> and may be removed in the future. Please use the new key
> >>>>>>> 'spark.kryoserializer.buffer' instead.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO Executor: Starting executor ID driver on
> host
> >>>>>>> localhost
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO Utils: Successfully started service
> >>>>>>> 'org.apache.spark.network.netty.NettyBlockTransferService' on port
> >>>> 64065.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO NettyBlockTransferService: Server created on
> >>>> 64065
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO BlockManagerMaster: Trying to register
> >>>>>> BlockManager
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO BlockManagerMasterEndpoint: Registering
> block
> >>>>>>> manager localhost:64065 with 2.4 GB RAM, BlockManagerId(driver,
> >>>>>> localhost,
> >>>>>>> 64065)
> >>>>>>>
> >>>>>>> 16/04/14 10:53:12 INFO BlockManagerMaster: Registered BlockManager
> >>>>>>>
> >>>>>>> java.lang.reflect.InvocationTargetException
> >>>>>>>
> >>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> >>>>>>>
> >>>>>>> at java.lang.reflect.Method.invoke(Method.java:497)
> >>>>>>>
> >>>>>>> at
> >>>>>>
> >> org.xerial.snappy.SnappyLoader.loadNativeLibrary(SnappyLoader.java:317)
> >>>>>>>
> >>>>>>> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:219)
> >>>>>>>
> >>>>>>> at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:169)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version$lzycompute(CompressionCodec.scala:168)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version(CompressionCodec.scala:168)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec.<init>(CompressionCodec.scala:152)
> >>>>>>>
> >>>>>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> >>>> Method)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> >>>>>>>
> >>>>>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65)
> >>>>>>>
> >>>>>>> at org.apache.spark.broadcast.TorrentBroadcast.org
> >>>>>>>
> >>>>>>
> >>>>
> >>
> $apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:73)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:80)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63)
> >>>>>>>
> >>>>>>> at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1326)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1014)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1011)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
> >>>>>>>
> >>>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714)
> >>>>>>>
> >>>>>>> at
> org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:1011)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:832)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:830)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
> >>>>>>>
> >>>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714)
> >>>>>>>
> >>>>>>> at org.apache.spark.SparkContext.textFile(SparkContext.scala:830)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.TDIndexedDatasetReader$class.elementReader(TextDelimitedReaderWriter.scala:61)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.math.indexeddataset.Reader$class.readElementsFrom(ReaderWriter.scala:75)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.readElementsFrom(TextDelimitedReaderWriter.scala:304)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:382)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:39)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.math.indexeddataset.package$.indexedDatasetDFSReadElements(package.scala:336)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$.readIndexedDatasets(ItemSimilarityDriver.scala:152)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$.process(ItemSimilarityDriver.scala:201)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:112)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:110)
> >>>>>>>
> >>>>>>> at scala.Option.map(Option.scala:145)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$.main(ItemSimilarityDriver.scala:110)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver.main(ItemSimilarityDriver.scala)
> >>>>>>>
> >>>>>>> Caused by: java.lang.UnsatisfiedLinkError: no snappyjava in
> >>>>>>> java.library.path
> >>>>>>>
> >>>>>>> at java.lang.ClassLoader.loadLibrary(ClassLoader.java:1864)
> >>>>>>>
> >>>>>>> at java.lang.Runtime.loadLibrary0(Runtime.java:870)
> >>>>>>>
> >>>>>>> at java.lang.System.loadLibrary(System.java:1122)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.xerial.snappy.SnappyNativeLoader.loadLibrary(SnappyNativeLoader.java:52)
> >>>>>>>
> >>>>>>> ... 49 more
> >>>>>>>
> >>>>>>> Exception in thread "main"
> >> java.lang.reflect.InvocationTargetException
> >>>>>>>
> >>>>>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> >>>> Method)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> >>>>>>>
> >>>>>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:72)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.CompressionCodec$.createCodec(CompressionCodec.scala:65)
> >>>>>>>
> >>>>>>> at org.apache.spark.broadcast.TorrentBroadcast.org
> >>>>>>>
> >>>>>>
> >>>>
> >>
> $apache$spark$broadcast$TorrentBroadcast$$setConf(TorrentBroadcast.scala:73)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:80)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:63)
> >>>>>>>
> >>>>>>> at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1326)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1014)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.SparkContext$$anonfun$hadoopFile$1.apply(SparkContext.scala:1011)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
> >>>>>>>
> >>>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714)
> >>>>>>>
> >>>>>>> at
> org.apache.spark.SparkContext.hadoopFile(SparkContext.scala:1011)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:832)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.SparkContext$$anonfun$textFile$1.apply(SparkContext.scala:830)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:150)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:111)
> >>>>>>>
> >>>>>>> at org.apache.spark.SparkContext.withScope(SparkContext.scala:714)
> >>>>>>>
> >>>>>>> at org.apache.spark.SparkContext.textFile(SparkContext.scala:830)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.TDIndexedDatasetReader$class.elementReader(TextDelimitedReaderWriter.scala:61)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.elementReader(TextDelimitedReaderWriter.scala:304)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.math.indexeddataset.Reader$class.readElementsFrom(ReaderWriter.scala:75)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.TextDelimitedIndexedDatasetReader.readElementsFrom(TextDelimitedReaderWriter.scala:304)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:382)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.sparkbindings.SparkEngine$.indexedDatasetDFSReadElements(SparkEngine.scala:39)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.math.indexeddataset.package$.indexedDatasetDFSReadElements(package.scala:336)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$.readIndexedDatasets(ItemSimilarityDriver.scala:152)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$.process(ItemSimilarityDriver.scala:201)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:112)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$$anonfun$main$1.apply(ItemSimilarityDriver.scala:110)
> >>>>>>>
> >>>>>>> at scala.Option.map(Option.scala:145)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver$.main(ItemSimilarityDriver.scala:110)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.mahout.drivers.ItemSimilarityDriver.main(ItemSimilarityDriver.scala)
> >>>>>>>
> >>>>>>> Caused by: java.lang.IllegalArgumentException:
> >>>>>>> org.xerial.snappy.SnappyError: [FAILED_TO_LOAD_NATIVE_LIBRARY] null
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:171)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version$lzycompute(CompressionCodec.scala:168)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec$.org$apache$spark$io$SnappyCompressionCodec$$version(CompressionCodec.scala:168)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec.<init>(CompressionCodec.scala:152)
> >>>>>>>
> >>>>>>> ... 38 more
> >>>>>>>
> >>>>>>> Caused by: org.xerial.snappy.SnappyError:
> >>>> [FAILED_TO_LOAD_NATIVE_LIBRARY]
> >>>>>>> null
> >>>>>>>
> >>>>>>> at org.xerial.snappy.SnappyLoader.load(SnappyLoader.java:229)
> >>>>>>>
> >>>>>>> at org.xerial.snappy.Snappy.<clinit>(Snappy.java:44)
> >>>>>>>
> >>>>>>> at
> >>>>>>>
> >>>>>>
> >>>>
> >>
> org.apache.spark.io.SnappyCompressionCodec$.liftedTree1$1(CompressionCodec.scala:169)
> >>>>>>>
> >>>>>>> ... 41 more
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO SparkContext: Invoking stop() from shutdown
> >> hook
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/metrics/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/stages/stage/kill,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/api,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/static,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/executors/threadDump/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/executors/threadDump,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/executors/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/executors,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/environment/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/environment,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/storage/rdd/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/storage/rdd,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/storage/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/storage,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/stages/pool/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/stages/pool,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/stages/stage/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/stages/stage,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/stages/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/stages,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/jobs/job/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/jobs/job,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/jobs/json,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ContextHandler: stopped
> >>>>>>> o.s.j.s.ServletContextHandler{/jobs,null}
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO SparkUI: Stopped Spark web UI at
> >>>>>>> http://192.168.1.122:4040
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO MapOutputTrackerMasterEndpoint:
> >>>>>>> MapOutputTrackerMasterEndpoint stopped!
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO MemoryStore: MemoryStore cleared
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO BlockManager: BlockManager stopped
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO BlockManagerMaster: BlockManagerMaster
> stopped
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO
> >>>>>>> OutputCommitCoordinator$OutputCommitCoordinatorEndpoint:
> >>>>>>> OutputCommitCoordinator stopped!
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO SparkContext: Successfully stopped
> >> SparkContext
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Shutdown hook called
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Deleting directory
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b/httpd-11f32419-5f8a-4b61-9654-9ea2d32cab5d
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO RemoteActorRefProvider$RemotingTerminator:
> >>>>>> Shutting
> >>>>>>> down remote daemon.
> >>>>>>>
> >>>>>>> 16/04/14 10:53:13 INFO ShutdownHookManager: Deleting directory
> >>>>>>>
> >>>>>>
> >>>>
> >>
> /private/var/folders/17/vhrqnx4x1n10yqbl1vnch9000000gn/T/spark-ad6a714b-6dcb-410c-aef3-2b098695743b
> >>>>>>>
> >>>>>>> --
> >>>>>>> Thanks & Regards,
> >>>>>>>
> >>>>>>> *Rohit Jain*
> >>>>>>> Web developer | Consultant
> >>>>>>> Mob +91 8097283931
> >>>>>>
> >>>>>>
> >>>>>
> >>>>>
> >>>>> --
> >>>>> Thanks & Regards,
> >>>>>
> >>>>> *Rohit Jain*
> >>>>> Web developer | Consultant
> >>>>> Mob +91 8097283931
> >>>>
> >>>>
> >>>
> >>>
> >>> --
> >>> Thanks & Regards,
> >>>
> >>> *Rohit Jain*
> >>> Web developer | Consultant
> >>> Mob +91 8097283931
> >>
> >>
> >
> >
> > --
> > Thanks & Regards,
> >
> > *Rohit Jain*
> > Web developer | Consultant
> > Mob +91 8097283931
>
>


-- 
Thanks & Regards,

*Rohit Jain*
Web developer | Consultant
Mob +91 8097283931

Reply via email to