If you want more parallelism, you need more cores. So, use a machine with more cores, or use a cluster of machines. spark-ec2<https://spark.apache.org/docs/latest/ec2-scripts.html>is the easiest way to do this.
If you're stuck on a single machine with 2 cores, then set your default parallelism to 2. Setting it to a higher number won't do anything helpful. On Fri, Apr 4, 2014 at 2:47 PM, Eduardo Costa Alfaia <e.costaalf...@unibs.it > wrote: > What do you advice me Nicholas? > > Em 4/4/14, 19:05, Nicholas Chammas escreveu: > > If you're running on one machine with 2 cores, I believe all you can get > out of it are 2 concurrent tasks at any one time. So setting your default > parallelism to 20 won't help. > > > On Fri, Apr 4, 2014 at 11:41 AM, Eduardo Costa Alfaia < > e.costaalf...@unibs.it> wrote: > >> Hi all, >> >> I have put this line in my spark-env.sh: >> -Dspark.default.parallelism=20 >> >> this parallelism level, is it correct? >> The machine's processor is a dual core. >> >> Thanks >> >> -- >> Informativa sulla Privacy: http://www.unibs.it/node/8155 >> > > > > Informativa sulla Privacy: http://www.unibs.it/node/8155 >