On 19/12/15 18:02, David Moreau Simard wrote: > The bulk of the jobs are offloaded to an ephemeral bare metal node but > ansible still runs from the slave. Some jobs do run locally, such as > smaller tox jobs. > > What's the strategy if we want to scale this ? At the moment the slaves all run from VM's hosted near the admin and jenkins instances - and are setup manually, and managed manually - this was very much a stop gap arrangement till we can get a better virtualised setup in place. We've been looking at and trying to scope up getting an RDO cloud in place, that could then be used for 3 things : 1) making an openstack api available for people who want to just consume VM's for their workloads 2) migrating the slaves into openstack managed ( ie, self/user managed ) virtual machines 3) offering up image backed resources for people looking at doing testing with other OSs, eg what the libvirt and libguestfs folks do at the moment. We have a dedicated hardware slab ( ~ 24 phy machines worth ) dedicated to this task ( so as to not cut into the ci baremetal pools ), but are waiting on the RH facility folks to get it wired up and dial-toned. Given the nature and impact of this setup, I am going to try and see if we can speed up delivery of that infra from the present timeline of end Feb '16. regards, -- Karanbir Singh +44-207-0999389 | http://www.karan.org/ | twitter.com/kbsingh GnuPG Key : http://www.karan.org/publickey.asc