Thanks Richard, On Wed, Apr 20, 2022 at 9:37 AM Richard Zowalla <[email protected]> wrote:
> Hi, > > seems we are the "top" consumers with 1,6TB disk usage on the CI > infrastructure. > > I looked at some of our jobs and found, that there is no retention > policy in place (for some of them). I added a policy similar to what we > had in the past for newly created jobs. Looks like the retention policy > is not copied then cloning jobs. > > In addition, I asked Gavin, if he can provide a "du" listing for our > jobs, so we can better dig into this issue. > Here is your listing: 834G master-deploy of which: 445G org.apache.tomee$apache-tomee 111G org.apache.tomee$tomee-embedded 60G org.apache.tomee$openejb-standalone 44G org.apache.tomee$tomee-plume-webapp 39G org.apache.tomee$tomee-plus-webapp 36G org.apache.tomee$tomee-microprofile-webapp 26G org.apache.tomee$tomee-webapp 25G org.apache.tomee$openejb-lite 5.4G org.apache.tomee$tomee-webaccess 5.1G org.apache.tomee$taglibs-shade 4.4G org.apache.tomee$openejb-provisionning 4.0G org.apache.tomee$openejb-itests-app 3.4G org.apache.tomee$openejb-ssh 3.3G org.apache.tomee$arquillian-tomee-moviefun-example 3.1G org.apache.tomee$cxf-shade 2.3G org.apache.tomee$openejb-core 597G jakarta-deploy of which: 354G org.apache.tomee$apache-tomee 73G org.apache.tomee$tomee-plume-webapp 66G org.apache.tomee$tomee-plus-webapp 60G org.apache.tomee$tomee-microprofile-webapp 42G org.apache.tomee$tomee-webprofile-webapp 4.3G org.apache.tomee$jakartaee-api 65M org.apache.tomee$tomee-project 44M org.apache.tomee$tomee 38M org.apache.tomee$transform 20M org.apache.tomee.jakarta$apache-tomee 7.2M org.apache.tomee.bom$jaxb-runtime 6.5M org.apache.tomee.bom$boms 63G tomee-8.x-deploy 25G jakarta-wip-9.x-deploy 23G master-build-full 7.0G site-publish 4.2G tomee-8.x-sanity-checks 3.1G tomee-7.0.x 2.5G master-sanity-checks 2.1G pull-request 2.0G tomee-8.x-build-full 2.0G TOMEE-3872 1.7G master-pull-request 1.1G tomee-8.x-owasp-check 1.1G master-owasp-check 1.1G master-build-quick 945M tomee-8.x-build-quick 35M tomee-jakartaee-api-master 27M tomee-patch-plugin-deploy 428K clean-repo 256K tomee-jenkins-pipelines > Gruß > Richard > > -------- Ursprüngliche Nachricht -------- > Von: Gavin McDonald <[email protected]> > Antwort an: [email protected], [email protected] > An: builds <[email protected]> > Betreff: ci-builds all 3.6TB disk is full! > Datum: Wed, 20 Apr 2022 09:27:28 +0200 > > > Hi All, > > > > Seems we need to do another cull of projects storing way too much > > data. > > > > Below are everyone above 1GB. Just FYI, 1GB is fine, likely 50GB is > > fine, > > but above > > that, its just too much. I will be removing 1TB of data from wherever > > I can > > get it. > > > > Please, look after your jobs, and your fellow projects by limiting > > what you > > keep. > > > > 1.6T Tomee > > 451G Kafka > > 303G james > > 176G carbondata > > 129G Jackrabbit > > 71G Brooklyn > > 64G Sling > > 64G Netbeans > > 60G Ranger > > 38G AsterixDB > > 33G OODT > > 29G Tika > > 27G Syncope > > 24G Atlas > > 20G IoTDB > > 18G CXF > > 16G POI > > 11G Solr > > 11G Mesos > > 8.7G Royale > > 7.8G Lucene > > 7.6G MyFaces > > 7.6G Directory > > 6.4G OpenJPA > > 6.0G ManifoldCF > > 5.9G ActiveMQ > > 5.7G Logging > > 5.6G Archiva > > 5.5G UIMA > > 5.3G ctakes > > 4.7G Heron > > 4.6G Jena > > 4.0G OpenOffice > > 3.8G Cloudstack > > 3.4G Shiro > > 2.5G Qpid > > 2.1G JSPWiki > > 2.1G JMeter > > 2.0G JClouds > > 1.8G Santuario > > 1.8G OpenMeetings > > 1.8G Camel > > 1.7G Karaf > > 1.7G HttpComponents > > 1.7G Ant > > 1.5G Tapestry > > 1.5G Commons > > 1.3G DeltaSpike > > 1.2G Rya > > 1.2G Aries > > 1.2G Accumulo > > 1.1G PDFBox > > > > -- > > > > *Gavin McDonald* > > Systems Administrator > > ASF Infrastructure Team > > -- *Gavin McDonald* Systems Administrator ASF Infrastructure Team
