Thanks for the listing. Retention policies are now in place for our daily deploy jobs, so we shouldn't produce this huge amount of data over time now.
If the retention kicks in, we should be fine. Gruß Richard Am Mittwoch, dem 20.04.2022 um 09:41 +0200 schrieb Gavin McDonald: > Thanks Richard, > > On Wed, Apr 20, 2022 at 9:37 AM Richard Zowalla <r...@apache.org> > wrote: > > Hi, > > > > seems we are the "top" consumers with 1,6TB disk usage on the CI > > infrastructure. > > > > I looked at some of our jobs and found, that there is no retention > > policy in place (for some of them). I added a policy similar to > > what we > > had in the past for newly created jobs. Looks like the retention > > policy > > is not copied then cloning jobs. > > > > In addition, I asked Gavin, if he can provide a "du" listing for > > our > > jobs, so we can better dig into this issue. > > Here is your listing: > > 834G master-deploy > > of which: > > 445G org.apache.tomee$apache-tomee > 111G org.apache.tomee$tomee-embedded > 60G org.apache.tomee$openejb-standalone > 44G org.apache.tomee$tomee-plume-webapp > 39G org.apache.tomee$tomee-plus-webapp > 36G org.apache.tomee$tomee-microprofile-webapp > 26G org.apache.tomee$tomee-webapp > 25G org.apache.tomee$openejb-lite > 5.4G org.apache.tomee$tomee-webaccess > 5.1G org.apache.tomee$taglibs-shade > 4.4G org.apache.tomee$openejb-provisionning > 4.0G org.apache.tomee$openejb-itests-app > 3.4G org.apache.tomee$openejb-ssh > 3.3G org.apache.tomee$arquillian-tomee-moviefun-example > 3.1G org.apache.tomee$cxf-shade > 2.3G org.apache.tomee$openejb-core > > 597G jakarta-deploy > > of which: > > 354G org.apache.tomee$apache-tomee > 73G org.apache.tomee$tomee-plume-webapp > 66G org.apache.tomee$tomee-plus-webapp > 60G org.apache.tomee$tomee-microprofile-webapp > 42G org.apache.tomee$tomee-webprofile-webapp > 4.3G org.apache.tomee$jakartaee-api > 65M org.apache.tomee$tomee-project > 44M org.apache.tomee$tomee > 38M org.apache.tomee$transform > 20M org.apache.tomee.jakarta$apache-tomee > 7.2M org.apache.tomee.bom$jaxb-runtime > 6.5M org.apache.tomee.bom$boms > > 63G tomee-8.x-deploy > 25G jakarta-wip-9.x-deploy > 23G master-build-full > 7.0G site-publish > 4.2G tomee-8.x-sanity-checks > 3.1G tomee-7.0.x > 2.5G master-sanity-checks > 2.1G pull-request > 2.0G tomee-8.x-build-full > 2.0G TOMEE-3872 > 1.7G master-pull-request > 1.1G tomee-8.x-owasp-check > 1.1G master-owasp-check > 1.1G master-build-quick > 945M tomee-8.x-build-quick > 35M tomee-jakartaee-api-master > 27M tomee-patch-plugin-deploy > 428K clean-repo > 256K tomee-jenkins-pipelines > > > Gruß > > Richard > > > > -------- Ursprüngliche Nachricht -------- > > Von: Gavin McDonald <gmcdon...@apache.org> > > Antwort an: bui...@apache.org, gmcdon...@apache.org > > An: builds <bui...@apache.org> > > Betreff: ci-builds all 3.6TB disk is full! > > Datum: Wed, 20 Apr 2022 09:27:28 +0200 > > > > > Hi All, > > > > > > Seems we need to do another cull of projects storing way too much > > > data. > > > > > > Below are everyone above 1GB. Just FYI, 1GB is fine, likely 50GB > > is > > > fine, > > > but above > > > that, its just too much. I will be removing 1TB of data from > > wherever > > > I can > > > get it. > > > > > > Please, look after your jobs, and your fellow projects by > > limiting > > > what you > > > keep. > > > > > > 1.6T Tomee > > > 451G Kafka > > > 303G james > > > 176G carbondata > > > 129G Jackrabbit > > > 71G Brooklyn > > > 64G Sling > > > 64G Netbeans > > > 60G Ranger > > > 38G AsterixDB > > > 33G OODT > > > 29G Tika > > > 27G Syncope > > > 24G Atlas > > > 20G IoTDB > > > 18G CXF > > > 16G POI > > > 11G Solr > > > 11G Mesos > > > 8.7G Royale > > > 7.8G Lucene > > > 7.6G MyFaces > > > 7.6G Directory > > > 6.4G OpenJPA > > > 6.0G ManifoldCF > > > 5.9G ActiveMQ > > > 5.7G Logging > > > 5.6G Archiva > > > 5.5G UIMA > > > 5.3G ctakes > > > 4.7G Heron > > > 4.6G Jena > > > 4.0G OpenOffice > > > 3.8G Cloudstack > > > 3.4G Shiro > > > 2.5G Qpid > > > 2.1G JSPWiki > > > 2.1G JMeter > > > 2.0G JClouds > > > 1.8G Santuario > > > 1.8G OpenMeetings > > > 1.8G Camel > > > 1.7G Karaf > > > 1.7G HttpComponents > > > 1.7G Ant > > > 1.5G Tapestry > > > 1.5G Commons > > > 1.3G DeltaSpike > > > 1.2G Rya > > > 1.2G Aries > > > 1.2G Accumulo > > > 1.1G PDFBox > > > > > > -- > > > > > > *Gavin McDonald* > > > Systems Administrator > > > ASF Infrastructure Team > > > >
signature.asc
Description: This is a digitally signed message part