@ Jillur Qudus aka Scammer

I know you are hiding on this mailing or at least or your friends are.

@Sean Owen Book/Theatre critic is a profession


When I first saw the following code on the introductory Page https://spark.apache.org/examples.htm

def inside(p):
    x, y = random.random(), random.random()
    return x*x + y*y < 1

count = sc.parallelize(xrange(0, NUM_SAMPLES)) \
             .filter(inside).count()
print "Pi is roughly %f" % (4.0 * count / NUM_SAMPLES)

I was a bit overwhelmed and thought WTF I better get a book which will explain this thoroughly. By  investing in a book I would get a thorough walk through of real cases and any dependency tools, with Spark  a Framework technology.

 

The Book I purchased was yours. Machine Learning with Apache Spark Quick Start Guide.

Www.Packt.com publishing. Birmingham – Mumbai

Author: Jillur Quddus

 

You describe yourself as “

lead technical architect polygot software engineer and data scientist with over ten years experience. Architecting and engineering distributed scalable high performance and secure solutions used to combat organized crime,cybercrime and fraud. .. working with central Government Intelligence , law enforcement , banking. Worked across the world including in Japan , Singapore , Malaysia , Hong Kong and New Zealand. Founder of UK based company Keisan specialising in open source distributed technologies and machine learning ..

 

on Page 53 you describe the step of installing ANACONDA { python package manager and python environment }

> bash Anaconda3-…

 

on Page 54

 

> conda install -c conda-forge pyspark

 

Page 63

You will need to restart any Jupyter Notebook instances, and the underlying Terminal Sessions

from which they were spawned, in order for SPARK_HOME environment variable to be successfully recognised and registered by findspark.”

 

I know that Spark Framework is a self contained framework with its own configurations files

so there is no need to set environment variables. If I wanted to use those SPARK_HOME libraries then the code reads like so.

 

import findspark

findspark.init('/path/to/spark-2.4.5-bin-hadoop2.7')

 

 

Page [63]

.. We are now ready to write our first spark Program in python ! .. it may be easier to split the following code

 

This is your code

 

# (1) import required Python Dependencies

import findspark

findspark.init(()

 

This will not compile without the slashes because this is PYTHON.

# (2) Instantiate the spark context

conf = SparkConf()

.setMaster(“spark://192.168.56.10:7077”)

.setAppName(“Calculate Pi”)

sc = SparkContext(conf=conf)

 

 

conf = SparkConf()\

.setMaster("spark://192.168.56.10:7077")\

.setAppName("Calculate Pi")\

sc = SparkContext(conf=conf)

 

 

This will compile because it is copied and pasted from Python example code.

# (3) Calculate the value of Pi i.e. 3.14 …

def inside (p):

x,y = random.random(), random.random()

return x*x + y*y <1

 

num_sample = 100

 

I have set it like so to get a more accurate value of pi

num_samples = 10**8

 

even this guy https://www.sicara.ai/blog/2017-05-02-get-started-pyspark-jupyter-notebook-3-minutes has set it to

num_samples = 100000000

"(I bet you understand what it does!)"


although good code would be num_samples = 10**8 (to the power of )


You also write in the book “To my wife and best friend Jane , for making life worth living. And to the memory of parents, for their sacrifices and giving me the freedom to explore my imagination.”

I understand these books' sale price is in India for $0.25 cents
because they printed there.

You obviously did not respect my freedom to explore and study  
and other who also spent their hard earning money and were SCAMMED like me.   

So I wish you a long life. In that long life I hope you have an accident where you are

parallelised from neck down and live the remainder of your years as a vegetables like Steven Hawkings. You wife has to spoon feed then finally end up and leaving you.

 

--------------------------------------------------------------------- To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to