Hi Lokesh,

Thanks for your advice.
I changed the heap size from 3GB to 4Gb and then to 8GB, set to
dfs.ratis.client.request.retry to 10 and decreased the number of pipeline
from 5 to 2.
It allows 10-15% better performance but it's still 35% slower than HDFS. I
also notice that the data is less well spread over the 4 nodes with only 2
pipelines.

I had a look at the GC log during the execution of the teragen (copy attach
this email), I'm not an expert at all in gc tuning so I load it the GC log
on this site and it tell me that for the 4 minutes of the execution of the
job the datanode was in pause during 29 seconds and trigger 1260 GC over 4
minutes. That looks pretty high for me but maybe not?
http://gcloganalyzer.com/?file=70338453-4505-4238-a900-c0206a2d52f4test.gc

I also take care that the yarn job don't request more than 80% of the VM
resource (even try with 60% but didn't change anything)

Do you think there is something else I can do to improve it or should I
stop here? Is it possible that the difference comes from the short circuit
feature of HDFS?

Thanks a lot,
Michel

Le mer. 29 juil. 2020 à 10:54, Lokesh Jain <lj...@apache.org> a écrit :

> Hi Michel
>
> Thanks for trying out Ozone!
> For ozone 0.5 - Can you please try another run after increasing config
> value for dfs.ratis.client.request.retry.interval to 10 or 15 seconds? The
> default value for this config is 1 second.
> > ozone.datanode.pipeline.limit
> Can you try a smaller value like 1 or 2 for above config with a data node
> heap size of 4GB? Please check GC pressure on the datanode with this config.
>
> There are some improvements which have gone recently after ozone 0.5
> release. I would also recommend to try the latest ozone.
>
> Thanks
> Lokesh
>
>
> > On 29-Jul-2020, at 12:57 AM, Michel Sumbul <michelsum...@gmail.com>
> wrote:
> >
> > I forgot to mention that I set ozone.datanode.pipeline.limit to 5.
> > Michel
> >
> > Le mar. 28 juil. 2020 à 20:22, Michel Sumbul <michelsum...@gmail.com
> <mailto:michelsum...@gmail.com>> a écrit :
> > Hi guys,
> >
> > I would like to know if you have any advice tips/tricks to get the best
> performance of Ozone? (Memory tuning / thread / specific settings / etc..)
> >
> > I did a few teragen/terasort on it and the results are really surprising
> compared to HDFS,Ozone (using the hadoopFS)  is almost 2 times slower than
> HDFS.
> >
> >
> >
> >
> >
> >
> > The clusters were exactly the same for both:
> > - 3 masters and 4 slaves (8core/32GB) (its a small cluster but that
> should matter here)
> > - Backend storage is a CEPH storage (80 servers)
> > - NIC: 2 x 25Gb/S
> > - Ozone version 0.5
> > - Each job was executed 5 times
> >
> > HDFS and Ozone were installed on the same nodes, one was down where the
> other one was up, to guarantee no other differences of configuration that
> the distributed FS.
> >
> > I was not expecting a big difference like this, do you have any idea how
> to improve this?
> > Or what can be the reason for that? I saw a few jira regarding data
> locality at read, can it be linked to that?
> >
> > Thanks,
> > Michel
> >
>
>
---------------------------------------------------------------------
To unsubscribe, e-mail: ozone-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: ozone-dev-h...@hadoop.apache.org

Reply via email to