1551-1600 of 10000 results (92ms)
2024-06-04 ยง
19:09 <marostegui@cumin1002> END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 1 day, 0:00:00 on db1241.eqiad.wmnet with reason: Maintenance [production]
19:09 <marostegui@cumin1002> START - Cookbook sre.hosts.downtime for 1 day, 0:00:00 on db1241.eqiad.wmnet with reason: Maintenance [production]
19:09 <marostegui@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db1238 (T364069)', diff saved to https://phabricator.wikimedia.org/P64034 and previous config saved to /var/cache/conftool/dbconfig/20240604-190906-marostegui.json [production]
19:06 <ryankemper> [WDQS Deploy] Restarting `wdqs-categories` across lvs-managed hosts, one node at a time: `sudo -E cumin -b 1 'A:wdqs-all and not A:wdqs-test' 'depool && sleep 45 && systemctl restart wdqs-categories && sleep 45 && pool'` [production]
19:06 <ryankemper> [WDQS Deploy] Restarted `wdqs-categories` across all test hosts simultaneously: `sudo -E cumin 'A:wdqs-test' 'systemctl restart wdqs-categories'` [production]
19:06 <ryankemper> [WDQS Deploy] Restarted `wdqs-updater` across all hosts, 4 hosts at a time: `sudo -E cumin -b 4 'A:wdqs-all' 'systemctl restart wdqs-updater'` [production]
19:00 <ryankemper@deploy1002> Finished deploy [wdqs/wdqs@43b966f]: 0.3.142 (duration: 12m 53s) [production]
18:53 <marostegui@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db1238', diff saved to https://phabricator.wikimedia.org/P64033 and previous config saved to /var/cache/conftool/dbconfig/20240604-185358-marostegui.json [production]
18:48 <ryankemper> [WDQS Deploy] Forgot to run the command to set git hash to tip of origin/master so deploy was a partial no-op. Re-rolling... [production]
18:47 <ryankemper@deploy1002> Started deploy [wdqs/wdqs@43b966f]: 0.3.142 [production]
18:46 <ryankemper@deploy1002> Finished deploy [wdqs/wdqs@143ca33]: 0.3.142 (duration: 02m 02s) [production]
18:45 <ryankemper> [WDQS Deploy] Tests passing following deploy of `0.3.142` on canary `wdqs1016`; proceeding to rest of fleet [production]
18:44 <ryankemper@deploy1002> Started deploy [wdqs/wdqs@143ca33]: 0.3.142 [production]
18:41 <ryankemper> [WDQS Deploy] Gearing up for deploy of wdqs `0.3.142`. Pre-deploy tests passing on canary `wdqs1016` [production]
18:38 <marostegui@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db1238', diff saved to https://phabricator.wikimedia.org/P64032 and previous config saved to /var/cache/conftool/dbconfig/20240604-183850-marostegui.json [production]
18:35 <mutante> aphlict - (phab realtime notifications) - reboots [production]
18:30 <mutante> doc.wikimedia.org - very short downtime for maintenance [production]
18:28 <dzahn@cumin1002> END (FAIL) - Cookbook sre.hosts.downtime (exit_code=99) for 0:10:00 on doc1003.eqiad.wmnet with reason: reboot T366555 [production]
18:28 <dzahn@cumin1002> START - Cookbook sre.hosts.downtime for 0:10:00 on doc1003.eqiad.wmnet with reason: reboot T366555 [production]
18:28 <dzahn@cumin1002> END (FAIL) - Cookbook sre.hosts.downtime (exit_code=99) for 0:10:00 on doc.wikimedia.org with reason: reboot T366555 [production]
18:28 <dzahn@cumin1002> START - Cookbook sre.hosts.downtime for 0:10:00 on doc.wikimedia.org with reason: reboot T366555 [production]
18:26 <dduvall@deploy1002> rebuilt and synchronized wikiversions files: group0 wikis to 1.43.0-wmf.8 refs T361402 [production]
18:23 <marostegui@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db1238 (T364069)', diff saved to https://phabricator.wikimedia.org/P64031 and previous config saved to /var/cache/conftool/dbconfig/20240604-182342-marostegui.json [production]
18:15 <kamila@cumin1002> END (FAIL) - Cookbook sre.hosts.reimage (exit_code=99) for host wikikube-ctrl1001.eqiad.wmnet with OS bullseye [production]
18:04 <sukhe@cumin1002> END (PASS) - Cookbook sre.cdn.roll-reboot (exit_code=0) rolling reboot on P{cp7014*} and A:cp [production]
17:54 <sukhe@cumin1002> START - Cookbook sre.cdn.roll-reboot rolling reboot on P{cp7014*} and A:cp [production]
17:53 <sukhe> sudo cumin 'A:cp-upload and A:magru' "sed -i '/\sup ethtool -A eno12399np0/d' /etc/network/interfaces" [production]
17:51 <sukhe> sudo cumin 'A:cp-text and A:magru' "sed -i '/\sup ethtool -A eno12399np0/d' /etc/network/interfaces" [production]
17:49 <sukhe@cumin1002> END (PASS) - Cookbook sre.cdn.roll-reboot (exit_code=0) rolling reboot on P{cp7002*} and A:cp [production]
17:39 <sukhe@cumin1002> START - Cookbook sre.cdn.roll-reboot rolling reboot on P{cp7002*} and A:cp [production]
17:23 <kamila@cumin1002> START - Cookbook sre.hosts.reimage for host wikikube-ctrl1001.eqiad.wmnet with OS bullseye [production]
17:22 <sukhe> sudo cumin 'A:cp and A:magru' 'run-puppet-agent' [production]
17:15 <kamila@cumin1002> END (PASS) - Cookbook sre.dns.netbox (exit_code=0) [production]
17:15 <kamila@cumin1002> END (PASS) - Cookbook sre.puppet.sync-netbox-hiera (exit_code=0) generate netbox hiera data: "Triggered by cookbooks.sre.dns.netbox: Moved wikikube-ctrl1001 to a new rack - kamila@cumin1002" [production]
17:14 <kamila@cumin1002> START - Cookbook sre.puppet.sync-netbox-hiera generate netbox hiera data: "Triggered by cookbooks.sre.dns.netbox: Moved wikikube-ctrl1001 to a new rack - kamila@cumin1002" [production]
17:11 <kamila@cumin1002> START - Cookbook sre.dns.netbox [production]
16:53 <sukhe@puppetmaster1001> conftool action : set/pooled=yes; selector: name=cp700[12].magru.wmnet,service=(cdn|ats-be) [production]
16:52 <swfrench@deploy1002> helmfile [eqiad] DONE helmfile.d/services/changeprop-jobqueue: apply [production]
16:51 <swfrench@deploy1002> helmfile [eqiad] START helmfile.d/services/changeprop-jobqueue: apply [production]
16:41 <elukey> delete other 2 pods in eventgate-main on wikikube-eqiad to test if envoy on them is in a weird state [production]
16:36 <dcaro@cumin1002> END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) for host cloudcephosd1010.eqiad.wmnet [production]
16:31 <elukey> delete 3 pods in eventgate-main on wikikube-eqiad to test if envoy on them is in a weird state [production]
16:29 <dcaro@cumin1002> START - Cookbook sre.hosts.reboot-single for host cloudcephosd1010.eqiad.wmnet [production]
16:22 <marostegui@cumin1002> dbctl commit (dc=all): 'db2203 (re)pooling @ 100%: Repooling', diff saved to https://phabricator.wikimedia.org/P64028 and previous config saved to /var/cache/conftool/dbconfig/20240604-162241-root.json [production]
16:22 <fabfur@cumin1002> END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) for host cp7002.magru.wmnet [production]
16:15 <fabfur@cumin1002> END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) for host cp7001.magru.wmnet [production]
16:12 <marostegui@cumin1002> dbctl commit (dc=all): 'Depooling db2137 (T364299)', diff saved to https://phabricator.wikimedia.org/P64025 and previous config saved to /var/cache/conftool/dbconfig/20240604-161233-marostegui.json [production]
16:12 <marostegui@cumin1002> END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 6:00:00 on db2137.codfw.wmnet with reason: Maintenance [production]
16:12 <marostegui@cumin1002> START - Cookbook sre.hosts.downtime for 6:00:00 on db2137.codfw.wmnet with reason: Maintenance [production]
16:12 <marostegui@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db2136 (T364299)', diff saved to https://phabricator.wikimedia.org/P64024 and previous config saved to /var/cache/conftool/dbconfig/20240604-161210-marostegui.json [production]