[
https://issues.apache.org/jira/browse/SYSTEMML-2396?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16513060#comment-16513060
]
LI Guobao edited comment on SYSTEMML-2396 at 6/14/18 10:14 PM:
---------------------------------------------------------------
[~mboehm7], is it correct this description? And also, after the last call, I
have thought about the solution that you mentioned which uses the array and
slot. However, I could not figure out what does the solution look like. Does it
mean that we launch an another pre-fetch thread per worker which will produce
the sliced matrices and put them into a cache?
was (Author: guobao):
[~mboehm7], is it correct this description? And also, after the last call, I
have thought about the solution that you mentioned which uses the array and
slot. However, I could not figure out what does the solution look like. Does it
mean that we launch an another pre-fetch thread per worker which will produce
the sliced matrices?
> Batch pre-fetching per workers
> ------------------------------
>
> Key: SYSTEMML-2396
> URL: https://issues.apache.org/jira/browse/SYSTEMML-2396
> Project: SystemML
> Issue Type: Sub-task
> Reporter: LI Guobao
> Assignee: LI Guobao
> Priority: Major
>
> This task aims to improve the performance of workers. Currently, in each
> iteration of mini-batch, we need to slice the matrix, execute the gradients
> computation and then send them to the ps for updating the model. While the ps
> is doing the aggregation work, the worker pauses due to waiting for the new
> model. Hence the idea is to completely use this free slot to pre-fetch the
> mini-batch in order to accelerate the future iteration.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)