2024-04-30
ยง
|
10:53 |
<jmm@cumin2002> |
END (PASS) - Cookbook sre.puppet.migrate-host (exit_code=0) for host db2189.codfw.wmnet |
[production] |
10:51 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'db1158 (re)pooling @ 25%: Repooling', diff saved to https://phabricator.wikimedia.org/P61492 and previous config saved to /var/cache/conftool/dbconfig/20240430-105124-root.json |
[production] |
10:48 |
<Dreamy_Jazz> |
Security deploy finished |
[production] |
10:47 |
<logmsgbot> |
dreamyjazz Deployed security patch for T338419 |
[production] |
10:45 |
<jmm@cumin2002> |
START - Cookbook sre.puppet.migrate-host for host db2189.codfw.wmnet |
[production] |
10:43 |
<jmm@cumin2002> |
START - Cookbook sre.hosts.reimage for host ganeti7001.magru.wmnet with OS bookworm |
[production] |
10:39 |
<aokoth@cumin1002> |
END (PASS) - Cookbook sre.hosts.decommission (exit_code=0) for hosts lists1004.eqiad.wmnet |
[production] |
10:39 |
<aokoth@cumin1002> |
END (PASS) - Cookbook sre.dns.netbox (exit_code=0) |
[production] |
10:39 |
<aokoth@cumin1002> |
END (PASS) - Cookbook sre.puppet.sync-netbox-hiera (exit_code=0) generate netbox hiera data: "Triggered by cookbooks.sre.dns.netbox: lists1004.eqiad.wmnet decommissioned, removing all IPs except the asset tag one - aokoth@cumin1002" |
[production] |
10:38 |
<jmm@cumin2002> |
END (PASS) - Cookbook sre.puppet.migrate-host (exit_code=0) for host db2175.codfw.wmnet |
[production] |
10:38 |
<aokoth@cumin1002> |
START - Cookbook sre.puppet.sync-netbox-hiera generate netbox hiera data: "Triggered by cookbooks.sre.dns.netbox: lists1004.eqiad.wmnet decommissioned, removing all IPs except the asset tag one - aokoth@cumin1002" |
[production] |
10:37 |
<btullis@cumin1002> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 2:00:00 on cephosd1001.eqiad.wmnet with reason: host reimage |
[production] |
10:36 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'db1158 (re)pooling @ 10%: Repooling', diff saved to https://phabricator.wikimedia.org/P61491 and previous config saved to /var/cache/conftool/dbconfig/20240430-103618-root.json |
[production] |
10:35 |
<btullis@cumin1002> |
START - Cookbook sre.hosts.downtime for 2:00:00 on cephosd1001.eqiad.wmnet with reason: host reimage |
[production] |
10:33 |
<logmsgbot> |
dreamyjazz Deployed security patch for T338419 |
[production] |
10:32 |
<aokoth@cumin1002> |
START - Cookbook sre.dns.netbox |
[production] |
10:27 |
<aokoth@cumin1002> |
START - Cookbook sre.hosts.decommission for hosts lists1004.eqiad.wmnet |
[production] |
10:25 |
<jmm@cumin2002> |
START - Cookbook sre.puppet.migrate-host for host db2175.codfw.wmnet |
[production] |
10:24 |
<jmm@cumin2002> |
END (PASS) - Cookbook sre.puppet.migrate-host (exit_code=0) for host db2148.codfw.wmnet |
[production] |
10:21 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'db1158 (re)pooling @ 5%: Repooling', diff saved to https://phabricator.wikimedia.org/P61490 and previous config saved to /var/cache/conftool/dbconfig/20240430-102113-root.json |
[production] |
10:16 |
<btullis@cumin1002> |
START - Cookbook sre.hosts.reimage for host cephosd1001.eqiad.wmnet with OS bullseye |
[production] |
10:15 |
<jmm@cumin2002> |
START - Cookbook sre.puppet.migrate-host for host db2148.codfw.wmnet |
[production] |
10:08 |
<jmm@cumin2002> |
END (PASS) - Cookbook sre.puppet.migrate-host (exit_code=0) for host db2138.codfw.wmnet |
[production] |
10:06 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'db1158 (re)pooling @ 1%: Repooling', diff saved to https://phabricator.wikimedia.org/P61489 and previous config saved to /var/cache/conftool/dbconfig/20240430-100607-root.json |
[production] |
10:00 |
<jmm@cumin2002> |
START - Cookbook sre.puppet.migrate-host for host db2138.codfw.wmnet |
[production] |
09:58 |
<jmm@cumin2002> |
END (PASS) - Cookbook sre.puppet.migrate-host (exit_code=0) for host db2125.codfw.wmnet |
[production] |
09:57 |
<arnaudb@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2179 (T360332)', diff saved to https://phabricator.wikimedia.org/P61488 and previous config saved to /var/cache/conftool/dbconfig/20240430-095745-arnaudb.json |
[production] |
09:53 |
<jiji@deploy1002> |
helmfile [codfw] DONE helmfile.d/admin 'sync'. |
[production] |
09:51 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'Push es6 codfw config T355424', diff saved to https://phabricator.wikimedia.org/P61487 and previous config saved to /var/cache/conftool/dbconfig/20240430-095119-marostegui.json |
[production] |
09:47 |
<jmm@cumin2002> |
START - Cookbook sre.puppet.migrate-host for host db2125.codfw.wmnet |
[production] |
09:46 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'Push es6 eqiad section T355285', diff saved to https://phabricator.wikimedia.org/P61486 and previous config saved to /var/cache/conftool/dbconfig/20240430-094635-marostegui.json |
[production] |
09:42 |
<arnaudb@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2179', diff saved to https://phabricator.wikimedia.org/P61485 and previous config saved to /var/cache/conftool/dbconfig/20240430-094237-arnaudb.json |
[production] |
09:41 |
<jayme@cumin1002> |
conftool action : set/pooled=inactive; selector: name=mw2382.codfw.wmnet |
[production] |
09:39 |
<Dreamy_Jazz> |
Starting security deploy on tmux session |
[production] |
09:32 |
<marostegui@cumin1002> |
END (PASS) - Cookbook sre.hosts.reimage (exit_code=0) for host db1158.eqiad.wmnet with OS bookworm |
[production] |
09:30 |
<marostegui@deploy1002> |
Finished scap: Backport for [[gerrit:1025670|etcd.php: Add es6 (T355285 T355424)]] (duration: 15m 01s) |
[production] |
09:29 |
<jiji@deploy1002> |
helmfile [codfw] START helmfile.d/admin 'sync'. |
[production] |
09:29 |
<jiji@deploy1002> |
helmfile [codfw] DONE helmfile.d/admin 'apply'. |
[production] |
09:28 |
<jiji@deploy1002> |
helmfile [codfw] START helmfile.d/admin 'apply'. |
[production] |
09:28 |
<jayme@cumin1002> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 2 days, 0:00:00 on mw2382.codfw.wmnet with reason: Degraded RAID/storage controller issues |
[production] |
09:28 |
<jayme@cumin1002> |
START - Cookbook sre.hosts.downtime for 2 days, 0:00:00 on mw2382.codfw.wmnet with reason: Degraded RAID/storage controller issues |
[production] |
09:28 |
<jiji@deploy1002> |
helmfile [staging-codfw] DONE helmfile.d/admin 'apply'. |
[production] |
09:27 |
<jiji@deploy1002> |
helmfile [staging-codfw] START helmfile.d/admin 'apply'. |
[production] |
09:27 |
<jiji@deploy1002> |
helmfile [staging-eqiad] DONE helmfile.d/admin 'apply'. |
[production] |
09:27 |
<jiji@deploy1002> |
helmfile [staging-eqiad] START helmfile.d/admin 'apply'. |
[production] |
09:27 |
<arnaudb@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2179', diff saved to https://phabricator.wikimedia.org/P61484 and previous config saved to /var/cache/conftool/dbconfig/20240430-092729-arnaudb.json |
[production] |
09:26 |
<jayme> |
draining mw2382.codfw.wmnet - T362938 |
[production] |
09:22 |
<arnaudb@cumin1002> |
dbctl commit (dc=all): 'db2151 (re)pooling @ 100%: Post replag', diff saved to https://phabricator.wikimedia.org/P61483 and previous config saved to /var/cache/conftool/dbconfig/20240430-092230-arnaudb.json |
[production] |
09:18 |
<volans@cumin1002> |
END (PASS) - Cookbook sre.deploy.python-code (exit_code=0) netbox to netbox-dev2002.codfw.wmnet with reason: Update Netbox dependencies for netbox-next - volans@cumin1002 |
[production] |
09:17 |
<marostegui@deploy1002> |
marostegui: Continuing with sync |
[production] |