On Mon, 28 Nov 2022, JOHN URBAN via clamav-users wrote:
We are experiencing a large number of MPI jobs failing indicating
the fabric is unavailable when the scans are running. Early in the
investigation so not sure if locking, timing, response time or other
factors are involved, but I wanted to ask a quick gernal question to
see if this is a known issue with easy answers. If not, we will post
more detailed information as it is determined.
Not an issue that I am familiar with.
Are the MPI jobs related to clamav, or just running
on a system with clamav ?
Is clamav doing on-access analysis ? If so I wonder whether it is
attempting to access the same file, or worse same file-handle, for
each mpi thread, simultaneously.
If I remember correctly "fabric" can be a technical term to do with
message passing, parallelism and networking.
Is that how you are using it ?
--
Andrew C. Aitchison Kendal, UK
[email protected]
_______________________________________________
Manage your clamav-users mailing list subscription / unsubscribe:
https://lists.clamav.net/mailman/listinfo/clamav-users
Help us build a comprehensive ClamAV guide:
https://github.com/Cisco-Talos/clamav-documentation
https://docs.clamav.net/#mailing-lists-and-chat