2020-06-29
§
|
08:03 |
<godog> |
prometheus eqiad -- lvextend --resizefs --size +200G vg-ssd/prometheus-ops |
[production] |
08:02 |
<marostegui@cumin1001> |
dbctl commit (dc=all): 'Slowly pool db1135 into s1 T253217', diff saved to https://phabricator.wikimedia.org/P11685 and previous config saved to /var/cache/conftool/dbconfig/20200629-080253-marostegui.json |
[production] |
07:46 |
<marostegui@cumin1001> |
dbctl commit (dc=all): 'Add db1135 (depooled) to s1 T253217', diff saved to https://phabricator.wikimedia.org/P11684 and previous config saved to /var/cache/conftool/dbconfig/20200629-074611-marostegui.json |
[production] |
07:16 |
<XioNoX> |
push new pfw firewall rules - T256170 |
[production] |
07:13 |
<marostegui> |
Deploy schema change on db1085 with replication to labs T253276 |
[production] |
07:12 |
<marostegui@cumin1001> |
dbctl commit (dc=all): 'Depool db1085', diff saved to https://phabricator.wikimedia.org/P11683 and previous config saved to /var/cache/conftool/dbconfig/20200629-071236-marostegui.json |
[production] |
06:53 |
<marostegui@cumin1001> |
dbctl commit (dc=all): 'Remove db1080 from MW', diff saved to https://phabricator.wikimedia.org/P11682 and previous config saved to /var/cache/conftool/dbconfig/20200629-065335-marostegui.json |
[production] |
06:50 |
<elukey> |
execute gnt-instance remove an-launcher1001.eqiad.wmnet on ganeti1011 - T256363 |
[production] |
06:47 |
<elukey@cumin1001> |
END (PASS) - Cookbook sre.hosts.decommission (exit_code=0) |
[production] |
06:46 |
<elukey@cumin1001> |
START - Cookbook sre.hosts.decommission |
[production] |
06:45 |
<marostegui> |
Deploy MCR schema change on db1090:3312 |
[production] |
06:35 |
<elukey> |
force puppet run on ores* to overcome celery OOMs on some nodes |
[production] |
04:57 |
<marostegui> |
Stop MySQL on db1080 to clone db1135 T253217 |
[production] |
04:56 |
<marostegui@cumin1001> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) |
[production] |
04:53 |
<marostegui@cumin1001> |
START - Cookbook sre.hosts.downtime |
[production] |
2020-06-27
§
|
20:22 |
<qchris> |
Gerrit upgrade done. |
[production] |
19:49 |
<mutante> |
removed 2620:0:861:3:208:80:154:136 from /etc/network/interfaces on gerrit1001, rebooting |
[production] |
19:27 |
<mutante> |
rebooting gerrit1001 one more time |
[production] |
19:24 |
<mutante> |
restarted ferm on gerrit1001 |
[production] |
19:19 |
<mutante> |
rebooting gerrit1001 one more time |
[production] |
19:05 |
<mutante> |
rebooting gerrit1001 |
[production] |
18:58 |
<mutante> |
rebooting gerrit2001 |
[production] |
18:49 |
<hashar> |
Enabling beta cluster update job (gerrit maintenance) https://integration.wikimedia.org/ci/view/Beta/job/beta-code-update-eqiad/ |
[production] |
18:35 |
<qchris@deploy1001> |
Finished deploy [gerrit/gerrit@da40615]: Gerrit to v3.2.2-98-g98d827eaa3 on gerrit2001 (duration: 00m 10s) |
[production] |
18:34 |
<qchris@deploy1001> |
Started deploy [gerrit/gerrit@da40615]: Gerrit to v3.2.2-98-g98d827eaa3 on gerrit2001 |
[production] |
18:27 |
<qchris@deploy1001> |
Finished deploy [gerrit/gerrit@da40615]: Gerrit to v3.2.2-98-g98d827eaa3 on gerrit1001 (duration: 00m 08s) |
[production] |
18:27 |
<qchris@deploy1001> |
Started deploy [gerrit/gerrit@da40615]: Gerrit to v3.2.2-98-g98d827eaa3 on gerrit1001 |
[production] |
17:25 |
<hashar> |
Disabled beta cluster update job (gerrit maintenance) https://integration.wikimedia.org/ci/view/Beta/job/beta-code-update-eqiad/ |
[production] |
17:19 |
<qchris> |
Stopping gerrit on gerrit1001 for the Gerrit upgrade |
[production] |
17:14 |
<qchris> |
Duplicating reviewdb changes so we get a cheap and quick rollback |
[production] |
17:11 |
<dzahn@cumin1001> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) |
[production] |
17:11 |
<dzahn@cumin1001> |
START - Cookbook sre.hosts.downtime |
[production] |
17:11 |
<qchris> |
Disabling puppet on gerrit1001 for Gerrit upgrades + data migrations |
[production] |
17:11 |
<dzahn@cumin1001> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) |
[production] |
17:11 |
<dzahn@cumin1001> |
START - Cookbook sre.hosts.downtime |
[production] |
17:07 |
<qchris> |
Starting Gerrit upgrade to v3.2.2-98-g98d827eaa3 |
[production] |
15:44 |
<qchris@deploy1001> |
Finished deploy [gerrit/gerrit@da40615]: Gerrit to v3.2.2-98-g98d827eaa3 on gerrit1002 (gerrit-test) (duration: 00m 08s) |
[production] |
15:44 |
<qchris@deploy1001> |
Started deploy [gerrit/gerrit@da40615]: Gerrit to v3.2.2-98-g98d827eaa3 on gerrit1002 (gerrit-test) |
[production] |
13:03 |
<qchris@deploy1001> |
Finished deploy [gerrit/gerrit@460e439]: Gerrit to v3.2.2-97-gcaf5020db1 on gerrit1002 (gerrit-test) (duration: 00m 08s) |
[production] |
13:03 |
<qchris@deploy1001> |
Started deploy [gerrit/gerrit@460e439]: Gerrit to v3.2.2-97-gcaf5020db1 on gerrit1002 (gerrit-test) |
[production] |
2020-06-26
§
|
18:42 |
<robh> |
all ulsfo onsite work completed as of 30 minutes ago |
[production] |
17:52 |
<robh> |
msw2-ulsfo work done, all mgmt items confirmed back online and icinga alerts cleared, moving onto msw1-ulsfo (rack 22) and will lose all mgmt in that rack for next 10-20 minutes T256300 |
[production] |
17:52 |
<robh> |
msw2-ulsfo work done, all mgmt items confirmed back online and icinga alerts cleared, moving onto msw1-ulsfo (rack 22) and will lose all mgmt in that rack for next 10-20 minutes |
[production] |
17:11 |
<robh> |
msw work in ulsfo via T256300 |
[production] |
10:24 |
<ema> |
pool 5006 T256449 |
[production] |
10:22 |
<marostegui@cumin1001> |
dbctl commit (dc=all): 'Repool db1085', diff saved to https://phabricator.wikimedia.org/P11677 and previous config saved to /var/cache/conftool/dbconfig/20200626-102248-marostegui.json |
[production] |
10:22 |
<marostegui@cumin1001> |
dbctl commit (dc=all): 'Repool db1093', diff saved to https://phabricator.wikimedia.org/P11676 and previous config saved to /var/cache/conftool/dbconfig/20200626-102201-marostegui.json |
[production] |
10:03 |
<ema> |
cp2039: restart purged T256444 |
[production] |