As previously announced in June (see
https://lists.centos.org/pipermail/ci-users/2022-June/004547.html and
also https://www.youtube.com/watch?v=aqgs-3NnRmA) we'll migrate
existing Duffy API service to the new Duffy v3, starting with Phase 1 .
Migration is scheduled for """"Monday August 1st 7:00 am UTC time"""".
You can convert to local time with $(date -d '2022-08-01 07:00 UTC')
The expected API downtime will be under the 60 seconds (DNS TTL).
Worth knowing:
- new duffy api (in legacy mode) has currently less available nodes, but
that will be resolved during the day (see below)
- we'll have to wait ~6h (default maximum "lease" time for a test node)
to be able to let new duffy api reinstall older seamicro nodes and so be
available in the new duffy v3 pool (transitioning from old to new duffy
pool)
- currently running jobs will continue so to work, but when you'll hit
new duffy api endpoint (DNS switch) to "return nodes" it will just
answer that session doesn't exist (new one) but it's safe to ignore
As soon as new duffy API is available you'll already be able to switch
your workflow to new the Duffy API and so use the new features.
Tenants/users documentation about how to interact with Duffy is
available on https://sigs.centos.org/guide/ci/ (see the Duffy part of
that documentation)
That means that switching to EC2 nodes will be directly possible.
During the day we'll tune the duffy nodepool configuration based on
usage metrics.
IMPORTANT remark : *only* projects/tenants that opted-in in the last 45
days are currently migrated (api and ssh keys), so if you haven't (yet),
you can still opt-in, otherwise all your requests for duffy ephemeral
nodes will be declined by new duffy api service next monday.
In case of troubles, both Pedro (nick phsmoura) and myself (nick arrfab)
will be also present in the #centos-ci irc channel on irc.libera.chat
the whole day.
Thanks for your understanding and patience.
on behalf of the Infra team,
--
Fabian Arrotin
The CentOS Project | https://www.centos.org
gpg key: 17F3B7A1 | twitter: @arrfab
Hello everyone,
This is a friendly reminder to opt-in if you are interested in using CentOS
CI for your project after the changes we announced last month (check
"[ci-users] Changes on CentOS CI and next steps
<https://lists.centos.org/pipermail/ci-users/2022-June/004547.html>").
*Projects
will be migrated only if they opt-in*. Please,* opt-in by August 2022*, you
can do that by simply replying to this email expressing your interest.
Also, we now have a deadline for the OpenShift new deployment as per our
last email. It should be up and running at the *beginning of September 2022*
.
Feel free to reach out if you have any questions or concerns,
--
Camila Granella
Associate Manager, Software Engineering
Community Platform Engineering
Red Hat <https://www.redhat.com/>
@Red Hat <https://twitter.com/redhat> Red Hat
<https://www.linkedin.com/company/red-hat> Red Hat
<https://www.facebook.com/RedHatInc>
<https://www.redhat.com/>
As you're all aware now, CentOS CI infra landscape will be different in
the next weeks/months.
One of the biggest change will be the introduction of Duffy v3 as
replacement for the Duffy API and while there will be compatibility
layer (allowing you to use your existing "cico" workflow), the goal is
that each tenant will move to the new api, which itself then requires
either:
- you abusing directly the api endpoint (https://duffy.ci.centos.org/docs)
- use the new duffy client (pip3.8 install --user duffy[client] with a
config file
We'd like to offer the duffy client natively in our cico-workspace
pod/container that is automatically deployed for your jenkins jobs on
openshift cluster.
But if we trigger it, it will also come with something additional : the
current centos Stream8 based container is using ansible 2.9.x, which is
now deprecated, and kicking a container rebuild would automatically
install ansible-core 2.12.x (present in el8 natively now)
As you're aware, ansible-core doesn't have all the modules that you
probably need to use for your workflow, so instead of install the
meta-pkg ansible (which then install almost everything) you can come
(reply to this thread) with a list of collections that you'd like to see
in the base container (the less we have to pull, the better, to keep the
container image as light as possible), we'd then add these to the
container image.
Thanks for your collaboration,
--
Fabian Arrotin
The CentOS Project | https://www.centos.org
gpg key: 17F3B7A1 | twitter: @arrfab
Hello!
After reading & discussing the recent news regarding the move to AWS, I wonder if it would be possible to provide Fedora Rawhide images along with the C8S and C9S ones (and maybe active stable Fedora releases as well).
A bit of background: In the systemd project we have several jobs with utilize Vagrant to run an Arch Linux VMs, in which we run tests alongside the C8S/C9S jobs, to cover issues with the latest-ish kernel and other software, and to also hunt down security issues with the latest versions of ASan and UBSan. However, all this is held together by a lot of duct tape and sheer will power, and in the end it requires an EC2 Metal instance to run, due to the additional level of virtualization.
If we were able to provision Rawhide instances directly (which should help us achieve the same goal as the Arch Linux VMs we currently use), that could, in theory, allow us to drop the requirement for Metal instances completely.
As far as I know, there are Fedora Rawhide AMIs[0], which should make this much easier, but that's all I can say, since I have almost zero experience with AWS overall.
Thank you!
Cheers,
Frantisek
[0] https://fedoraproject.org/w/index.php?title=Test_Results:Current_Cloud_Test…
--
PGP Key ID: 0xFB738CE27B634E4B
Just a quick reminder that CentOS Linux 8 is going EOL in december (see
https://www.centos.org/centos-linux-eol/)
So based on that plan, we'll also remove it from available ci infra at
the same time as on mirror network, and if you'll ask for CentOS 8
you'll get no returned node to run your tests on.
Start switching your tests to run on 8-stream instead, available since
the beginning.
--
Fabian Arrotin
The CentOS Project | https://www.centos.org
gpg key: 17F3B7A1 | twitter: @arrfab
Due to a storage migration (see
https://pagure.io/centos-infra/issue/534) we'll have to shutdown and
restart the openshift CI cluster (OCP)
Migration is scheduled for """"Tuesday January 18th, 9:00 am UTC time"""".
You can convert to local time with $(date -d '2022-01-18 09:00 UTC')
The expected "downtime" is estimated to ~60 minutes , time needed to :
- shutdown openshift workers and control plane nodes
- last data sync between old and new nfs storage node
- restart openshift control plane and workers nodes
Thanks for your comprehending and patience.
on behalf of the Infra team,
--
--
Fabian Arrotin
The CentOS Project | https://www.centos.org
gpg key: 17F3B7A1 | twitter: @arrfab
Hi all CI infra tenants/users,
As a follow-up wrt previous mail about cico-workspace container update
(see
https://lists.centos.org/pipermail/ci-users/2021-October/002184.html) we
had a look at what was needed to rebase the cico-workspace container to
CentOS 8-stream.
We've built a container that you can start using with the following specs:
- based on centos 8-stream
- so having python3 (instead of python2), important for your jobs
- still providing python-cicoclient (python3 version)
- ansible 2.9.27
- git
This container is actually available as staging tag, so it's not *yet*
the one that is automatically pulled through your jenkins as
"cloud/container" node to run your jobs.
We'd like you to start testing your jobs against that updated
cico-workspace, at your earliest convenience, so that you have a chance
to eventually adapt your jobs, or ask us to add/change something in that
container.
See the next items
# How to test ?
The current DeploymentConfig used in openshift will provide jenkins,
itself configure to pull cico-workspace from quay.io
(quay.io/centosci/cico-workspace:latest)
But you can also , as you have full admin rights in your own jenkins
namespace, temporary switch to :staging tag.
To proceed, just log into your own jenkins apps on openshift/ocp.ci and
follow the next steps:
Manage Jenkins
Manage Nodes and Clouds (under System Configuration)
Configure Clouds (on the left menu)
Expand Kubernetes and under Pod template you'll see "cico-workspace"
"Pod Template details"
Containers
Just swap quay.io/centosci/cico-workspace:latest =>
quay.io/centosci/cico-workspace:staging
Apply/Save
You can now test your job (test job) and see if that still works or if
you need to adapt something
# How to revert ?
Assuming that you want to revert to :latest test (still c7 based), just
proceed as above and swap back to quay.io/centosci/cico-workspace:latest
#When will you enforce the change ?
We have no fixed date in mind but we'd like to enforce this early
december, which would give you enough time to fix your jobs (if there is
a need to). We can also provide an alternative through a different tag
like :c7 while we'd have c8s for :latest and you'd then be also able to
modify your jobs to just continue to use older c7 container (not adviced
though, but we'll continue to rebuild the container for security updates
automatically)
Should you have questions/remarks, feel free to chime in !
--
Fabian Arrotin
The CentOS Project | https://www.centos.org
gpg key: 17F3B7A1 | twitter: @arrfab
For people not following inititial infra ticket about this
(https://pagure.io/centos-infra/issue/441) here is a quick status about
CentOS Stream 9 availability in CI infra :
Starting from today, each tenant in CI infra can request Stream 9 for
the x86_64 architecture.
Due to lack of compatible hardware for ppc64le, it's not available yet
but we hope to be able to provide it before end of this year.
aarch64 is on the list of architectures to be tested and added (more on
that when it will be ready)
We already updated the python-cicoclient to correct version on the
previous Virtual Machines setup behind ci.centos.org but we still have
(next on the list) to update the cico-workspace container/image to
include newer python-cicoclient (see
http://mirror.centos.org/centos/7/infra/x86_64/infra-common/Packages/p/pyth…)
that is used in the openshift/ocp 4 cluster
Enjoy your tests on Stream 9 !
PS: at the infra side, we'll still have to also have an internal mirror
returned by mirrormanager as it's actually using external mirrors, so
dnf operations will be "slow" versus what you can see for 7/8/8-stream,
but it's on the list
--
Fabian Arrotin
The CentOS Project | https://www.centos.org
gpg key: 17F3B7A1 | twitter: @arrfab
Follow-up on the previous mail about supporting 9-stream in CI infra.
We have refreshed the cico-workspace container that is deployed as
jenkins agent container and it's available on
https://quay.io/repository/centosci/cico-workspace
That's the version that is automatically deployed in the OCP cluster.
Changes:
- updated to CentOS 7.9 + updates
- updated python-cicoclient to 0.4.6 (version that supports requesting
for '9-stream' nodes to duffy)
- added ansible 2.9.27 (per request from some tenants, avoiding to
have to install it themselves)
Next steps: We'd like to rebase that container to 8-stream in the near
future but we'd just need first to have python-cicoclient available as
pkg and tested with python3. Once we'll have such test image, we'll
rebase and keep you informed about it
--
Fabian Arrotin
The CentOS Project | https://www.centos.org
gpg key: 17F3B7A1 | twitter: @arrfab
Due to some needed software updates, we'll have to shutdown and restart
jenkins on the 'legacy' ci.centos.org setup (reminder : moving to
openshift would be great and we'll have to put a hard deadline for this
if some tenants are still on the old setup)
Migration is scheduled for """"Tuesday September 28th, 9:00 am UTC time"""".
You can convert to local time with $(date -d '2021-09-28 09:00 UTC')
The expected "downtime" is estimated to ~30 minutes , time needed to
put jenkins in "shutdown/quiet mode" , wait for some some jobs to
finish, update and restart jenkins.
Thanks for your comprehending and patience.
on behalf of the Infra team,
--
Fabian Arrotin
The CentOS Project | https://www.centos.org
gpg key: 17F3B7A1 | twitter: @arrfab