2017-06-12
§
|
12:01 |
<marostegui@tin> |
Synchronized wmf-config/db-eqiad.php: Depool db1089 for maintenance - T166935 (duration: 00m 41s) |
[production] |
12:01 |
<moritzm> |
upgrading mw1266-mw1275 to HHVM 3.18 |
[production] |
11:09 |
<joal@tin> |
Finished deploy [analytics/refinery@d9c3419]: Regular weekly deploy of refinery (mostly unique_devices patches) (duration: 06m 18s) |
[production] |
11:05 |
<moritzm> |
upgrading job runners mw1162-mw1164 to HHVM 3.18 |
[production] |
11:03 |
<joal@tin> |
Started deploy [analytics/refinery@d9c3419]: Regular weekly deploy of refinery (mostly unique_devices patches) |
[production] |
10:59 |
<marostegui> |
Drop table updates on commonswiki (s4) - T139342 |
[production] |
10:28 |
<moritzm> |
upgrading mw1250-mw1258 to HHVM 3.18 |
[production] |
09:55 |
<moritzm> |
upgrading mw1221-mw1235 to HHVM 3.18 |
[production] |
09:25 |
<godog> |
swift eqiad-prod finish decom ms-be1005/6/7 - T166489 |
[production] |
09:13 |
<moritzm> |
upgrading mw1236-mw1249 to HHVM 3.18 |
[production] |
09:12 |
<marostegui> |
Drop table updates on dewiki and wikidatawiki (s5) - T139342 |
[production] |
08:31 |
<godog> |
reboot ms-be1002, load avg slowly creeping up |
[production] |
08:22 |
<elukey> |
powercycle scb2005 (console frozen, host unresponsive) |
[production] |
07:40 |
<elukey> |
restarted citoid on scb1001 (kept failing health checks for Error: write EPIPE) |
[production] |
07:38 |
<marostegui> |
Reboot https://icinga.wikimedia.org/cgi-bin/icinga/extinfo.cgi?type=1&host=ms-be1008 as xfs is failing |
[production] |
07:31 |
<marostegui> |
Deploy alter table s2 - db1060 - T166205 |
[production] |
07:31 |
<marostegui@tin> |
Synchronized wmf-config/db-eqiad.php: Depool db1060 - T166205 (duration: 00m 41s) |
[production] |
07:26 |
<elukey> |
ran restart-pdfrender on scb1001 (OOM errors in the dmesg from hours ago) |
[production] |
07:22 |
<elukey> |
ran restart-pdfrender on scb1002 (OOM errors in the dmesg from hours ago) |
[production] |
07:21 |
<marostegui> |
Deploy alter table s4 - db1064 - https://phabricator.wikimedia.org/T166206 |
[production] |
07:19 |
<marostegui@tin> |
Synchronized wmf-config/db-eqiad.php: Depool db1064 - T166206 (duration: 00m 41s) |
[production] |
06:53 |
<moritzm> |
upgrade remaining app servers running HHVM 3.18 to 3.18.2+wmf5 |
[production] |
05:38 |
<marostegui> |
Deploy alter table s4 - labsdb1003 - T166206 |
[production] |
02:14 |
<l10nupdate@tin> |
scap failed: average error rate on 1/11 canaries increased by 10x (rerun with --force to override this check, see https://logstash.wikimedia.org/goto/3888cca979647b9381a7739b0bdbc88e for details) |
[production] |
2017-06-09
§
|
21:18 |
<mobrovac@tin> |
Finished deploy [restbase/deploy@4e5cb35]: (no justification provided) (duration: 01m 40s) |
[production] |
21:17 |
<mobrovac@tin> |
Started deploy [restbase/deploy@4e5cb35]: (no justification provided) |
[production] |
21:07 |
<mobrovac@tin> |
Finished deploy [restbase/deploy@4e5cb35]: Ensure the extract field is always present in the summary response - T167045 (take #2) (duration: 05m 23s) |
[production] |
21:02 |
<mobrovac@tin> |
Started deploy [restbase/deploy@4e5cb35]: Ensure the extract field is always present in the summary response - T167045 (take #2) |
[production] |
21:01 |
<mobrovac@tin> |
Finished deploy [restbase/deploy@4e5cb35]: Ensure the extract field is always present in the summary response - T167045 (duration: 04m 57s) |
[production] |
20:56 |
<mobrovac@tin> |
Started deploy [restbase/deploy@4e5cb35]: Ensure the extract field is always present in the summary response - T167045 |
[production] |
20:54 |
<mobrovac@tin> |
Finished deploy [restbase/deploy@4e5cb35] (staging): Ensure the extract field is always present in the summary response (duration: 03m 39s) |
[production] |
20:50 |
<mobrovac@tin> |
Started deploy [restbase/deploy@4e5cb35] (staging): Ensure the extract field is always present in the summary response |
[production] |
20:12 |
<demon@tin> |
Synchronized php-1.30.0-wmf.4/extensions/CirrusSearch/includes/Job/DeleteArchive.php: Really fix it this time (duration: 00m 43s) |
[production] |
19:49 |
<mutante> |
fermium: $ sudo /usr/local/sbin/disable_list wikino-bureaucrats (T166848) |
[production] |
19:46 |
<RainbowSprinkles> |
mw1299: running scap pull, maybe out of date? |
[production] |
18:12 |
<gehel> |
retry allocation of failed shards on elasticsearch eqiad |
[production] |
15:47 |
<_joe_> |
installed python-service-checker 0.1.3 on einsteinium,tegmen T167048 |
[production] |
15:44 |
<_joe_> |
uploaded service-checker 0.1.3 |
[production] |
15:11 |
<_joe_> |
upgraded python-service-checker to 0.1.2 on tegmen,einsteinium |
[production] |
13:18 |
<godog> |
upgrade thumbor to 0.1.40 - T167462 |
[production] |
12:36 |
<gehel> |
reducing high watermark on elasticsearch eqiad to rebalance shards |
[production] |