Hi, > >> Looks like ASF Jenkins's lucene slave is running out of disk space (again): > >> > >>> [smoker] [junit4] java.io.IOException: No space left on device > >> > >> Uwe, opinions on what to do? Should we ask Infra for more disk space > >> (again)? > > I’m guessing you must have done the manual deletions? I don’t see any > orphaned workspaces on disk, and available space looks adequate:
We should also nuke the workspaces of the disabled jobs of 5.4! I am clicking through that in the GUI... Uwe > —— > sarowe@lucene1-us-west:~$ df /home/jenkins/jenkins-slave/workspace > Filesystem 1K-blocks Used Available Use% Mounted on > /dev/sdb1 82437808 56211492 22015680 72% /x1 > —— > > Disk footprints for each job’s workspace: > > —— > sarowe@lucene1-us-west:~$ sudo -u jenkins du -sk /home/jenkins/jenkins- > slave/workspace/* > 121436 /home/jenkins/jenkins-slave/workspace/Apache Jackrabbit Oak > matrix > 973492 /home/jenkins/jenkins-slave/workspace/Lucene-Artifacts-5.3 > 984084 /home/jenkins/jenkins-slave/workspace/Lucene-Artifacts-5.4 > 980732 /home/jenkins/jenkins-slave/workspace/Lucene-Artifacts-5.x > 963896 /home/jenkins/jenkins-slave/workspace/Lucene-Artifacts-trunk > 6527688 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > Clover-5.x > 6131460 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > Clover-trunk > 1523640 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > Maven-5.3 > 1641880 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > Maven-5.4 > 1644900 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > Maven-5.x > 1651268 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > Maven-trunk > 2654308 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > NightlyTests-5.3 > 4435392 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > NightlyTests-5.4 > 2370868 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > NightlyTests-5.x > 3191700 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > NightlyTests-trunk > 2988924 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > SmokeRelease-5.3 > 1917096 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > SmokeRelease-5.4 > 1917780 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > SmokeRelease-5.x > 1919780 /home/jenkins/jenkins-slave/workspace/Lucene-Solr- > SmokeRelease-trunk > 1017100 /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests- > 5.3-Java7 > 1040176 /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests- > 5.4-Java7 > 1047236 /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests- > 5.x-Java7 > 988956 /home/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-trunk- > Java8 > 1618928 /home/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.3 > 1638788 /home/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.4 > 1641804 /home/jenkins/jenkins-slave/workspace/Solr-Artifacts-5.x > 1649436 /home/jenkins/jenkins-slave/workspace/Solr-Artifacts-trunk > —— > > Totals by branch: > > —— > 16496496 Trunk > 16131008 5.x > 11657416 5.4 > 10776392 5.3 > 55182748 Total > —— > > So it looks like as long as we keep up with pruning inactive jobs and orphaned > workspaces, we can safely accomodate two (or three?) active development > branches and two (or three?) active release branches with currently available > disk space. > > We’ll have to keep an eye on it when branch_6x and associated Jenkins jobs > are created. > > Steve > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: [email protected] > For additional commands, e-mail: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
