[ 
https://issues.apache.org/jira/browse/BEAM-14368?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Beam JIRA Bot updated BEAM-14368:
---------------------------------
    Labels: run-inference stale-assigned  (was: run-inference)

> Investigate load state_dict vs loading whole model
> --------------------------------------------------
>
>                 Key: BEAM-14368
>                 URL: https://issues.apache.org/jira/browse/BEAM-14368
>             Project: Beam
>          Issue Type: Sub-task
>          Components: sdk-py-core
>            Reporter: Anand Inguva
>            Assignee: Anand Inguva
>            Priority: P2
>              Labels: run-inference, stale-assigned
>          Time Spent: 2h 20m
>  Remaining Estimate: 0h
>
> Loading pytorch model as whole has some issues with pickling. Investigate it 
> with running some experiments. If the model size is too large, the current 
> implementation of the RunInference for PyTorch would fail because of memory 
> limits.
>  
> 1. We can pass the model class to the `load_model` of PyTorchModelLoader and 
> load the model there. This wouldn't pickle the model object but would pickle 
> the class and the model would be instantiated on the workers.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

Reply via email to