Maybe you can debug a little if you see that happen again? Perhaps something like this to get the largest number of instances, if there is a per-file leak:
IN: scratchpad all-instances [ class-of ] histogram-by sort-values reverse 10 head . Some other words for inspecting memory: http://docs.factorcode.org/content/article-tools.memory.html Can you give us some information about your disk layout? Is it one big directory with 1 million files? Is it a tree of directories? What do you think is average number of files per-directory? I opened a bug report if you'd like to provide feedback there rather than the mailing list: https://github.com/slavapestov/factor/issues/1483 On Thu, Oct 1, 2015 at 8:38 AM, HP wei <hpwe...@gmail.com> wrote: > Well, I just checked the running factor session that failed the > task overnight that I mentioned in below email. > > From the linux system command 'top', > I see that this particular factor is using > VIRT 4.0g > RES 2.0g > %MEM 26% > > I clicked on the restart listener button and the numbers remain the same. > should I have done more to clean up the memory usage ? > > ------------------ > > For comparison, I killed the factor session and restart it from the shell. > The numbers are > VIRT 940M > RES 182M > %MEM 2.2% > > ==> Had the factor continued to run last night, > it would have probably exhausted the memory on the machine. > I guess there might be some memory (leak) issue somewhere ??? > > --HP > > > > ---------- Forwarded message ---------- > From: HP wei <hpwe...@gmail.com> > Date: Thu, Oct 1, 2015 at 9:36 AM > Subject: how to error trapping 'link-info' > To: factor-talk@lists.sourceforge.net > > > As suggested by John, I test out the following action to > get the total file sizes of a disk volume. > > 0 "a_path_to_big_folder" [ link-info dup symbolic-link? [ drop ] [ size>> > + ] if ] each-file > > > Our big-folder is on a netapp server shared by tens of people. Many small > files get updated > every minutes if not seconds. The update may involve removing the file > first. > It has many many subfolders which in turn have more subfolders. > Each subfolder may have hundreds of files (occasionally in the thousands). > > After a few day's discussion with factor guru's, I understand that > each-file traverses the directory structure by first putting > entries of a folder in a sequence. And it processes each entry one by one. > Although this may not cause using big chunk of memory at a time, > it does have the following issue.. > > ------------------------ > > Last night, I left the command running and came back this morning to find > that it failed with the message. > lstat: "a path to a file" does not exist !!! > > This is because after 'each-file' puts the file into the sequence and then > when > it is its turn to be processed, it is not there at the time!! > Without error trapping, the above "0 ... each-file" could not work in our > case. > > So, I guess I would need to do error-trapping on the word link-info. > I do not know how to do it. Any hint ? > > Thanks > HP > > > > > ------------------------------------------------------------------------------ > > _______________________________________________ > Factor-talk mailing list > Factor-talk@lists.sourceforge.net > https://lists.sourceforge.net/lists/listinfo/factor-talk > >
------------------------------------------------------------------------------
_______________________________________________ Factor-talk mailing list Factor-talk@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/factor-talk