[Ci-users] No space left on device

Thu Dec 12 09:10:03 UTC 2019
Jaime Melis <jmelis at redhat.com>

I'm backing it up manually, so no need for step 2.

Can I ping you once the backup is done so you can put the slave in
maintenance mode?


On Thu, 12 Dec 2019 at 10:05, Vipul Siddharth <vipul at redhat.com> wrote:
>
> On Thu, Dec 12, 2019 at 2:32 PM Jaime Melis <jmelis at redhat.com> wrote:
> >
> > Hi Vipul,
> >
> > The slave is full. Basically too many workspaces with a fair amount of
> > data in them:
> >
> > 15.5 GiB /workspace
> >
> > And these are the top 10 dirs under /workspace:
> >
> > 547.9 MiB /devtools-openshift-jenkins-s2i-config-fabric8-push-prcheck at 2
> > 547.3 MiB /devtools-openshift-jenkins-s2i-config-fabric8-push-prcheck
> > 464.2 MiB /devtools-kubernetes-model at 2
> > 460.9 MiB /devtools-kubernetes-model-fabric8-push-build-master
> > 422.0 MiB /devtools-openshift-jenkins-s2i-config-fabric8-push-build-master
> > 362.3 MiB /devtools-kubernetes-client-fabric8-push-build-master
> > 241.8 MiB /devtools-kubernetes-client-fabric8-push-prcheck at 4
> > 236.0 MiB /devtools-kubernetes-client-fabric8-push-prcheck at 6
> > 236.0 MiB /devtools-kubernetes-client-fabric8-push-prcheck at 5
> > 206.5 MiB /devtools-contract-test-consumer-fabric8-wit-fabric8-auth
> >
> > So, I would propose to delete all those, in order to do that I would
> > like to request the following:
> >
> > 1. put the slave in maintenance mode
> > 2. take snapshot of the slave's disk
> > 3. we will delete all the workspaces
> > 4. re-enable the slave node
> >
> > Can you do 1, 2 and 4, and coordinate with me to do 3?
> Thank you Jaime for the help,
> I am looking into it
> Appreciate your help
> >
> > Thanks,
> > Jaime
> >
> >
> > On Thu, 12 Dec 2019 at 09:41, Vipul Siddharth <vipul at redhat.com> wrote:
> > >
> > > The Jenkins master has good space left, it could be that the slave
> > > node itself is full.
> > > Can you please share the build link?
> > >
> > > On Thu, Dec 12, 2019 at 1:53 PM Katerina Foniok <kkanova at redhat.com> wrote:
> > > >
> > > > Hello guys,
> > > >
> > > > from yesterday evening we are experiencing failures in our jobs on ci.centos.org saying No space left on device. It's blocking our PR checks.
> > > >
> > > > Could you please take a look at what's happening there? Thank you!
> > > >
> > > > Have a nice day,
> > > > Katka
> > > > _______________________________________________
> > > > Ci-users mailing list
> > > > Ci-users at centos.org
> > > > https://lists.centos.org/mailman/listinfo/ci-users
> > >
> > >
> > >
> > > --
> > > Vipul Siddharth
> > > He/His/Him
> > > Fedora | CentOS CI Infrastructure Team
> > > Red Hat
> > > w: vipul.dev
> > >
> > > _______________________________________________
> > > Ci-users mailing list
> > > Ci-users at centos.org
> > > https://lists.centos.org/mailman/listinfo/ci-users
> > >
> >
> >
> > --
> > Jaime Melis
> > Application SRE team, Service Delivery
> > Red Hat
> > jmelis at redhat.com
> >
>
>
> --
> Vipul Siddharth
> He/His/Him
> Fedora | CentOS CI Infrastructure Team
> Red Hat
> w: vipul.dev
>


--
Jaime Melis
Application SRE team, Service Delivery
Red Hat
jmelis at redhat.com