2021-01-14
§
|
00:20 |
<ryankemper@cumin1001> |
END (PASS) - Cookbook sre.elasticsearch.rolling-restart (exit_code=0) |
[production] |
00:15 |
<chaomodus> |
completed rebooting Netbox hosts, failure was due to report errors that would not have recovered. |
[production] |
00:14 |
<crusnov@cumin1001> |
END (FAIL) - Cookbook sre.hosts.reboot-single (exit_code=1) |
[production] |
00:13 |
<ryankemper> |
`sudo -i cookbook sre.elasticsearch.rolling-restart relforge "relforge cluster restart" --task-id T266492 --nodes-per-run 1 --without-lvs` |
[production] |
00:13 |
<ryankemper> |
(Forgot to tell it `relforge` isn't lvs-managed) |
[production] |
00:13 |
<ryankemper@cumin1001> |
START - Cookbook sre.elasticsearch.rolling-restart |
[production] |
00:10 |
<ryankemper@cumin1001> |
END (FAIL) - Cookbook sre.elasticsearch.rolling-restart (exit_code=99) |
[production] |
00:10 |
<ryankemper> |
T266492 Beginning rolling restart of `relforge` |
[production] |
00:09 |
<ryankemper@cumin1001> |
START - Cookbook sre.elasticsearch.rolling-restart |
[production] |
00:04 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2240.codfw.wmnet |
[production] |
00:04 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2239.codfw.wmnet |
[production] |
00:04 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2238.codfw.wmnet |
[production] |
00:04 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2237.codfw.wmnet |
[production] |
00:01 |
<crusnov@cumin1001> |
START - Cookbook sre.hosts.reboot-single |
[production] |
00:01 |
<crusnov@cumin1001> |
END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) |
[production] |
00:00 |
<ryankemper> |
T266492 T268779 T265699 Rolling restart of `cloudelastic` was successful |
[production] |
2021-01-13
§
|
23:53 |
<crusnov@cumin1001> |
START - Cookbook sre.hosts.reboot-single |
[production] |
23:53 |
<crusnov@cumin1001> |
END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) |
[production] |
23:49 |
<ryankemper@cumin1001> |
END (PASS) - Cookbook sre.elasticsearch.rolling-restart (exit_code=0) |
[production] |
23:49 |
<crusnov@cumin1001> |
START - Cookbook sre.hosts.reboot-single |
[production] |
23:49 |
<crusnov@cumin1001> |
END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) |
[production] |
23:46 |
<crusnov@cumin1001> |
START - Cookbook sre.hosts.reboot-single |
[production] |
23:46 |
<crusnov@cumin1001> |
END (FAIL) - Cookbook sre.hosts.reboot-single (exit_code=99) |
[production] |
23:46 |
<crusnov@cumin1001> |
START - Cookbook sre.hosts.reboot-single |
[production] |
23:44 |
<chaomodus> |
rebooting Netbox instances to apply updates |
[production] |
23:18 |
<dzahn@cumin1001> |
conftool action : set/pooled=no; selector: name=mw2240.codfw.wmnet |
[production] |
23:18 |
<dzahn@cumin1001> |
conftool action : set/pooled=no; selector: name=mw2239.codfw.wmnet |
[production] |
23:18 |
<dzahn@cumin1001> |
conftool action : set/pooled=no; selector: name=mw2238.codfw.wmnet |
[production] |
23:18 |
<dzahn@cumin1001> |
conftool action : set/pooled=no; selector: name=mw2237.codfw.wmnet |
[production] |
22:53 |
<ryankemper@cumin1001> |
START - Cookbook sre.elasticsearch.rolling-restart |
[production] |
22:53 |
<ryankemper> |
T266492 T268779 T265699 `sudo -i cookbook sre.elasticsearch.rolling-restart cloudelastic "cloudelastic cluster restart" --task-id T266492 --nodes-per-run 1` |
[production] |
22:53 |
<ryankemper> |
T266492 T268779 T265699 Restarting cloudelastic to apply new readahead changes, this will also verify cloudelastic support works in our elasticsearch spicerack code. Only going one node at a time because cloudelastic elasticsearch indices only have 1 replica shard per index |
[production] |
21:51 |
<dzahn@cumin1001> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 4:00:00 on mw2239.codfw.wmnet with reason: new install on buster |
[production] |
21:51 |
<dzahn@cumin1001> |
START - Cookbook sre.hosts.downtime for 4:00:00 on mw2239.codfw.wmnet with reason: new install on buster |
[production] |
21:44 |
<dzahn@cumin1001> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 2:00:00 on mw2240.codfw.wmnet with reason: REIMAGE |
[production] |
21:42 |
<dzahn@cumin1001> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 2:00:00 on mw2238.codfw.wmnet with reason: REIMAGE |
[production] |
21:41 |
<dzahn@cumin1001> |
START - Cookbook sre.hosts.downtime for 2:00:00 on mw2240.codfw.wmnet with reason: REIMAGE |
[production] |
21:40 |
<dzahn@cumin1001> |
END (FAIL) - Cookbook sre.hosts.downtime (exit_code=99) for 2:00:00 on mw2239.codfw.wmnet with reason: REIMAGE |
[production] |
21:40 |
<dzahn@cumin1001> |
END (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 2:00:00 on mw2237.codfw.wmnet with reason: REIMAGE |
[production] |
21:40 |
<dzahn@cumin1001> |
START - Cookbook sre.hosts.downtime for 2:00:00 on mw2239.codfw.wmnet with reason: REIMAGE |
[production] |
21:39 |
<dzahn@cumin1001> |
START - Cookbook sre.hosts.downtime for 2:00:00 on mw2238.codfw.wmnet with reason: REIMAGE |
[production] |
21:38 |
<dzahn@cumin1001> |
START - Cookbook sre.hosts.downtime for 2:00:00 on mw2237.codfw.wmnet with reason: REIMAGE |
[production] |
21:14 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2235.codfw.wmnet |
[production] |
21:14 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2234.codfw.wmnet |
[production] |
21:13 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2233.codfw.wmnet |
[production] |
21:12 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2232.codfw.wmnet |
[production] |
21:12 |
<dzahn@cumin1001> |
conftool action : set/pooled=yes; selector: name=mw2231.codfw.wmnet |
[production] |
21:06 |
<dzahn@cumin1001> |
conftool action : set/pooled=no; selector: name=mw2235.codfw.wmnet |
[production] |
21:06 |
<dzahn@cumin1001> |
conftool action : set/pooled=no; selector: name=mw2234.codfw.wmnet |
[production] |
21:06 |
<dzahn@cumin1001> |
conftool action : set/pooled=no; selector: name=mw2233.codfw.wmnet |
[production] |