9201-9250 of 10000 results (59ms)
2022-02-07 §
11:18 <hnowlan@deploy1002> helmfile [codfw] DONE helmfile.d/services/changeprop-jobqueue: sync on production [production]
11:18 <hnowlan@deploy1002> helmfile [codfw] DONE helmfile.d/services/changeprop-jobqueue: sync on staging [production]
11:18 <hnowlan@deploy1002> helmfile [codfw] START helmfile.d/services/changeprop-jobqueue: sync on production [production]
11:15 <hnowlan@deploy1002> helmfile [eqiad] DONE helmfile.d/services/changeprop-jobqueue: sync on production [production]
11:14 <hnowlan@deploy1002> helmfile [eqiad] DONE helmfile.d/services/changeprop-jobqueue: sync on staging [production]
11:14 <hnowlan@deploy1002> helmfile [eqiad] START helmfile.d/services/changeprop-jobqueue: sync on production [production]
11:03 <btullis> depooled aqs1006 [analytics]
11:00 <btullis@puppetmaster1001> conftool action : set/pooled=inactive; selector: name=aqs1006.eqiad.wmnet [production]
10:51 <mmandere> rolling upgrade of varnish from version 6.0.9 to 6.0.10 across DCs T300264 [production]
10:49 <filippo@puppetmaster1001> conftool action : set/pooled=no; selector: name=prometheus2004.codfw.wmnet [production]
10:49 <filippo@puppetmaster1001> conftool action : set/pooled=no; selector: name=prometheus1004.eqiad.wmnet [production]
10:22 <btullis@puppetmaster1001> conftool action : set/pooled=inactive; selector: name=aqs1005.eqiad.wmnet [production]
10:22 <btullis> depooling aqs1005 [analytics]
09:59 <btullis@puppetmaster1001> conftool action : set/pooled=inactive; selector: name=aqs1004.eqiad.wmnet [production]
09:21 <godog> temp-disable mfa for 'filippo' - T296629 [production]
09:09 <jayme> uncordoned kubernetes1014 - T301099 [production]
08:02 <jayme> powercycle kubernetes1014 - T301099 [production]
06:20 <jayme@cumin1001> END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 5 days, 0:00:00 on kubernetes1014.eqiad.wmnet with reason: potential HW error [production]
06:20 <jayme@cumin1001> START - Cookbook sre.hosts.downtime for 5 days, 0:00:00 on kubernetes1014.eqiad.wmnet with reason: potential HW error [production]
06:10 <jayme> draining kubernetes1014 [production]
2022-02-06 §
18:55 <wm-bot> <lucaswerkmeister> Checked out WSOAuth change Ie0a828e368 PS2 (commit c85a5a1c42), hopefully fixes T298351 [tools.notwikilambda]
18:36 <taavi> manually add T301081 fixes to prevent further k8s cluster issues [tools.mismatch-finder]
18:36 <taavi> manually add T301081 fixes to prevent further k8s cluster issues [tools.mismatch-finder-staging]
18:23 <taavi> deleted cronjob "mismatch-finder-staging.schedule" creating too many job objects [tools.mismatch-finder-staging]
2022-02-05 §
22:10 <andrew@cumin1001> END (PASS) - Cookbook sre.hosts.reimage (exit_code=0) for host cloudvirt2003-dev.codfw.wmnet with OS bullseye [production]
21:28 <andrew@cumin1001> START - Cookbook sre.hosts.reimage for host cloudvirt2003-dev.codfw.wmnet with OS bullseye [production]
20:15 <andrew@cumin1001> END (PASS) - Cookbook sre.hosts.reimage (exit_code=0) for host cloudvirt2002-dev.codfw.wmnet with OS bullseye [production]
19:29 <andrew@cumin1001> START - Cookbook sre.hosts.reimage for host cloudvirt2002-dev.codfw.wmnet with OS bullseye [production]
18:48 <andrew@cumin1001> END (PASS) - Cookbook sre.hosts.reimage (exit_code=0) for host cloudvirt2001-dev.codfw.wmnet with OS bullseye [production]
17:53 <andrew@cumin1001> START - Cookbook sre.hosts.reimage for host cloudvirt2001-dev.codfw.wmnet with OS bullseye [production]
16:54 <andrew@cumin1001> END (PASS) - Cookbook sre.hosts.reimage (exit_code=0) for host cloudvirt2001-dev.codfw.wmnet with OS bullseye [production]
06:11 <andrew@cumin1001> START - Cookbook sre.hosts.reimage for host cloudvirt2001-dev.codfw.wmnet with OS bullseye [production]
06:09 <andrew@cumin1001> END (FAIL) - Cookbook sre.hosts.reimage (exit_code=99) for host cloudvirt2001-dev.codfw.wmnet with OS bullseye [production]
05:41 <andrew@cumin1001> START - Cookbook sre.hosts.reimage for host cloudvirt2001-dev.codfw.wmnet with OS bullseye [production]
00:09 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [community-labs-monitoring]
00:09 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [antiharassment]
00:08 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [annotation]
00:07 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [wikidata-realtime-dumps]
00:06 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [wikidata-autodesc]
00:06 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [thumbor]
00:05 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [sentry]
00:02 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [ircwebchat]
00:01 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [graphql]
00:01 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [globalcu]
00:00 <andrewbogott> stopping all VMs as per https://wikitech.wikimedia.org/wiki/News/Cloud_VPS_2021_Purge [glampipe]
2022-02-04 §
23:59 <andrewbogott> accidentally restarted all VMs due to misreading the project purge page. sorry! [cvn]
23:43 <jhathaway@cumin1001> END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 1:00:00 on mirror1001.wikimedia.org with reason: new kernel [production]
23:43 <jhathaway@cumin1001> START - Cookbook sre.hosts.downtime for 1:00:00 on mirror1001.wikimedia.org with reason: new kernel [production]
23:02 <inflatador> bking@deployment-puppetmaster04 local commit to public/private repo, see T299797 for more details [production]
22:50 <wm-bot> <root> Cleaned up old log files [tools.ru_monuments]