2951-3000 of 10000 results (94ms)
2024-05-13 ยง
19:26 <cdanis@deploy1002> helmfile [staging] START helmfile.d/services/opentelemetry-collector: apply [production]
18:49 <herron@cumin1002> END (PASS) - Cookbook sre.kafka.roll-restart-reboot-brokers (exit_code=0) rolling restart_daemons on A:kafka-logging-eqiad [production]
18:49 <ryankemper@cumin2002> END (PASS) - Cookbook sre.kafka.roll-restart-mirror-maker (exit_code=0) restart MirrorMaker for Kafka A:kafka-mirror-maker-jumbo-eqiad cluster: Roll restart of jvm daemons. [production]
18:30 <ryankemper@cumin2002> START - Cookbook sre.kafka.roll-restart-mirror-maker restart MirrorMaker for Kafka A:kafka-mirror-maker-jumbo-eqiad cluster: Roll restart of jvm daemons. [production]
18:24 <herron@cumin1002> START - Cookbook sre.kafka.roll-restart-reboot-brokers rolling restart_daemons on A:kafka-logging-eqiad [production]
18:24 <ryankemper@cumin2002> END (PASS) - Cookbook sre.kafka.roll-restart-reboot-brokers (exit_code=0) rolling restart_daemons on A:kafka-jumbo-eqiad [production]
18:20 <swfrench@deploy1002> helmfile [eqiad] DONE helmfile.d/services/blubberoid: apply [production]
18:19 <swfrench@deploy1002> helmfile [eqiad] START helmfile.d/services/blubberoid: apply [production]
18:08 <swfrench@deploy1002> helmfile [codfw] DONE helmfile.d/services/blubberoid: apply [production]
18:07 <swfrench@deploy1002> helmfile [codfw] START helmfile.d/services/blubberoid: apply [production]
18:04 <swfrench@deploy1002> helmfile [staging] DONE helmfile.d/services/blubberoid: apply [production]
18:03 <swfrench@deploy1002> helmfile [staging] START helmfile.d/services/blubberoid: apply [production]
17:40 <cmooney@cumin1002> END (PASS) - Cookbook sre.dns.netbox (exit_code=0) [production]
17:40 <cmooney@cumin1002> END (PASS) - Cookbook sre.puppet.sync-netbox-hiera (exit_code=0) generate netbox hiera data: "Triggered by cookbooks.sre.dns.netbox: add records for new linknets on codfw spines - cmooney@cumin1002" [production]
17:39 <cmooney@cumin1002> START - Cookbook sre.puppet.sync-netbox-hiera generate netbox hiera data: "Triggered by cookbooks.sre.dns.netbox: add records for new linknets on codfw spines - cmooney@cumin1002" [production]
17:38 <ebernhardson@deploy1002> helmfile [eqiad] DONE helmfile.d/services/cirrus-streaming-updater: apply [production]
17:38 <ebernhardson@deploy1002> helmfile [eqiad] START helmfile.d/services/cirrus-streaming-updater: apply [production]
17:37 <cmooney@cumin1002> START - Cookbook sre.dns.netbox [production]
17:27 <ryankemper> T363973 [Kafka] Restarting `jumbo-eqiad` brokers, followed by mirror maker [production]
17:27 <ryankemper@cumin2002> START - Cookbook sre.kafka.roll-restart-reboot-brokers rolling restart_daemons on A:kafka-jumbo-eqiad [production]
17:05 <jayme@cumin1002> END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 2:00:00 on kubestagemaster2005.codfw.wmnet with reason: host reimage [production]
17:02 <jayme@cumin1002> START - Cookbook sre.hosts.downtime for 2:00:00 on kubestagemaster2005.codfw.wmnet with reason: host reimage [production]
16:50 <jayme@cumin1002> START - Cookbook sre.hosts.reimage for host kubestagemaster2005.codfw.wmnet with OS bullseye [production]
16:49 <jayme@cumin1002> END (PASS) - Cookbook sre.ganeti.changedisk (exit_code=0) for changing disk type of kubestagemaster2005.codfw.wmnet to plain [production]
16:47 <jayme@cumin1002> START - Cookbook sre.ganeti.changedisk for changing disk type of kubestagemaster2005.codfw.wmnet to plain [production]
16:47 <jayme@cumin1002> END (PASS) - Cookbook sre.ganeti.changedisk (exit_code=0) for changing disk type of kubestagemaster2004.codfw.wmnet to plain [production]
16:46 <jayme@cumin1002> START - Cookbook sre.ganeti.changedisk for changing disk type of kubestagemaster2004.codfw.wmnet to plain [production]
16:46 <ejegg> fundraising civicrm upgraded from c0d2fa95 to 4f55a7cf [production]
16:46 <jayme@cumin1002> END (FAIL) - Cookbook sre.ganeti.makevm (exit_code=97) for new host kubestagemaster2005.codfw.wmnet [production]
16:46 <jayme@cumin1002> END (ERROR) - Cookbook sre.hosts.reimage (exit_code=97) for host kubestagemaster2005.codfw.wmnet with OS bullseye [production]
16:34 <brouberol@cumin2002> END (PASS) - Cookbook sre.elasticsearch.rolling-operation (exit_code=0) Operation.RESTART (3 nodes at a time) for ElasticSearch cluster search_codfw: JVM restart - brouberol@cumin2002 - T363975 [production]
16:16 <jdrewniak@deploy1002> Synchronized portals: Wikimedia Portals Update: [[gerrit:1031006| Bumping portals to master (T128546)]] (duration: 13m 47s) [production]
16:13 <ejegg> restarted fundraising scheduled jobs [production]
16:11 <ejegg> fundraising civicrm rolled back from 3fef5849 to c0d2fa95 [production]
16:02 <jdrewniak@deploy1002> Synchronized portals/wikipedia.org/assets: Wikimedia Portals Update: [[gerrit:1031006| Bumping portals to master (T128546)]] (duration: 14m 23s) [production]
15:59 <ladsgroup@cumin1002> dbctl commit (dc=all): 'Depooling db1170 (T352010)', diff saved to https://phabricator.wikimedia.org/P62370 and previous config saved to /var/cache/conftool/dbconfig/20240513-155911-ladsgroup.json [production]
15:59 <ladsgroup@cumin1002> END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 1 day, 0:00:00 on db1170.eqiad.wmnet with reason: Maintenance [production]
15:58 <ladsgroup@cumin1002> START - Cookbook sre.hosts.downtime for 1 day, 0:00:00 on db1170.eqiad.wmnet with reason: Maintenance [production]
15:58 <ladsgroup@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db1158 (T352010)', diff saved to https://phabricator.wikimedia.org/P62369 and previous config saved to /var/cache/conftool/dbconfig/20240513-155849-ladsgroup.json [production]
15:55 <ejegg> fundraising civicrm upgraded from c0d2fa95 to 3fef5849 [production]
15:54 <ejegg> disabled fundraising scheduled jobs for CiviCRM deploy [production]
15:49 <herron@cumin1002> END (PASS) - Cookbook sre.kafka.roll-restart-reboot-brokers (exit_code=0) rolling restart_daemons on A:kafka-logging-codfw [production]
15:43 <ladsgroup@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db1158', diff saved to https://phabricator.wikimedia.org/P62368 and previous config saved to /var/cache/conftool/dbconfig/20240513-154341-ladsgroup.json [production]
15:28 <ladsgroup@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db1158', diff saved to https://phabricator.wikimedia.org/P62367 and previous config saved to /var/cache/conftool/dbconfig/20240513-152833-ladsgroup.json [production]
15:27 <elukey@deploy1002> helmfile [ml-staging-codfw] Ran 'sync' command on namespace 'experimental' for release 'main' . [production]
15:25 <herron@cumin1002> START - Cookbook sre.kafka.roll-restart-reboot-brokers rolling restart_daemons on A:kafka-logging-codfw [production]
15:19 <marostegui@cumin1002> END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 8:00:00 on db2098.codfw.wmnet with reason: Maintenance [production]
15:19 <marostegui@cumin1002> START - Cookbook sre.hosts.downtime for 8:00:00 on db2098.codfw.wmnet with reason: Maintenance [production]
15:18 <brouberol@cumin2002> START - Cookbook sre.elasticsearch.rolling-operation Operation.RESTART (3 nodes at a time) for ElasticSearch cluster search_codfw: JVM restart - brouberol@cumin2002 - T363975 [production]
15:13 <ladsgroup@cumin1002> dbctl commit (dc=all): 'Repooling after maintenance db1158 (T352010)', diff saved to https://phabricator.wikimedia.org/P62366 and previous config saved to /var/cache/conftool/dbconfig/20240513-151325-ladsgroup.json [production]