251-300 of 10000 results (42ms)
2018-06-12 ยง
21:39 <bblack> cp3043 - starting process to move to reimage into cache_upload [production]
19:16 <dduvall@deploy1001> rebuilt and synchronized wikiversions files: group0 to 1.32.0-wmf.8 [production]
18:57 <herron> restarted icinga service on einsteinium [production]
18:42 <dduvall@deploy1001> Finished scap: testwiki to php-1.32.0-wmf.8 and rebuild l10n cache (duration: 39m 39s) [production]
18:03 <dduvall@deploy1001> Started scap: testwiki to php-1.32.0-wmf.8 and rebuild l10n cache [production]
17:38 <ariel@deploy1001> Finished deploy [dumps/dumps@038c8b3]: sync after snapshot1009 install (duration: 00m 04s) [production]
17:37 <ariel@deploy1001> Started deploy [dumps/dumps@038c8b3]: sync after snapshot1009 install [production]
17:37 <ariel@deploy1001> Finished deploy [dumps/dumps@038c8b3]: sync after snapshot1009 install (duration: 00m 07s) [production]
17:37 <ariel@deploy1001> Started deploy [dumps/dumps@038c8b3]: sync after snapshot1009 install [production]
16:54 <marxarelli> starting branch cut for 1.32.0-wmf.8 [production]
16:11 <volans@deploy1001> Finished deploy [debmonitor/deploy@0eca14a]: Release v0.1.3 (duration: 00m 22s) [production]
16:11 <volans@deploy1001> Started deploy [debmonitor/deploy@0eca14a]: Release v0.1.3 [production]
15:40 <bblack> cp3034 - nevermind, doing different approach later in the day, still pooled in text for now! [production]
15:29 <bblack> cp3043 switching from text to upload shortly, downtimed in icinga for 2h - https://gerrit.wikimedia.org/r/c/operations/puppet/+/439936 [production]
15:07 <ema> cp3039: restart varnish-backend [production]
14:38 <addshore> file exporter importer slot done [production]
14:38 <addshore@deploy1001> Synchronized wmf-config/InitialiseSettings.php: FileImporter/Exporter [[gerrit:439876|Enable FileExporter/Importer on group0 wikis]] T195370 (duration: 00m 51s) [production]
14:20 <addshore@deploy1001> Synchronized wmf-config/CommonSettings.php: FileImporter/Exporter [[gerrit:439875|Allow setting of export target for FileExporter]] T195370 (duration: 00m 50s) [production]
14:09 <addshore@deploy1001> Finished scap: [[gerrit:439900|FileExporter backport]] - Pre deployment backport (extension not yet deployed) (duration: 30m 37s) [production]
13:38 <addshore@deploy1001> Started scap: [[gerrit:439900|FileExporter backport]] - Pre deployment backport (extension not yet deployed) [production]
13:16 <moritzm> installing openjdk-8 security updates on restbase-dev along with cassandra restarts [production]
12:38 <ema> cp3035: restart varnish-be, mbox lag [production]
12:34 <_joe_> repooling mw1230 after reimaging T196881 [production]
12:14 <marostegui> Deploy schema change on db1099:3311 T191316 T192926 T89737 T195193 [production]
12:14 <marostegui@deploy1001> Synchronized wmf-config/db-eqiad.php: Depool db1099:3311 for alter table (duration: 00m 52s) [production]
12:11 <marostegui> Deploy schema change on dbstore1002:s1 T191316 T192926 T89737 T195193 [production]
12:05 <akosiaris@puppetmaster1001> conftool action : set/weight=10; selector: dc=.*,service=mathoid,cluster=kubernetes,name=.* [production]
11:47 <moritzm> updated component/cassandra311 on apt.wikimedia.org to 3.11.2 [production]
10:26 <jynus> setting expire_log_days on db1066 as 30 [production]
10:21 <godog> bounce stuck rsyslog on lithium / wezen - T136312 [production]
09:41 <vgutierrez> cp3037 has been depooled due to unknown hardware issues T196974 [production]
08:48 <marostegui> Stop replication on db2094 to change triggers for archive table [production]
08:36 <volans> running puppet on failed hosts post small puppet outage and puppetdb reboot [production]
08:35 <akosiaris> rebalance ganeti codfw cluster [production]
08:35 <ema@neodymium> conftool action : set/pooled=no; selector: name=cp3037.esams.wmnet [production]
08:33 <akosiaris> reboot puppetdb1001 for spec-ctrl enable. Bundling it with a minor puppet outage to only have a torrent of harmless puppet failures once [production]
08:15 <akosiaris> ganeti2002 reboot for microcode update [production]
08:04 <akosiaris> ganeti2006 reboot for microcode update [production]
08:03 <marostegui> Deploy schema change on s1 codfw primary master (db2048) with replication, this will generate lag on codfw T191316 T192926 T89737 T195193 [production]
07:53 <marostegui@deploy1001> Synchronized wmf-config/db-eqiad.php: Repool db1121 after alter table (duration: 00m 50s) [production]
07:43 <akosiaris> ganeti2007 reboot for microcode update [production]
07:41 <akosiaris> ganeti2003 reboot for microcode update [production]
07:31 <mutante> closing idle screen session on tin (about to be decomed, dont use anymore) [production]
06:37 <marostegui> Deploy schema change on db1121 with replication, this will generate lag on labsdb:s4 T191316 T192926 T89737 T195193 [production]
06:37 <marostegui@deploy1001> Synchronized wmf-config/db-eqiad.php: Depool db1121 for alter table (duration: 00m 50s) [production]
06:31 <marostegui> Stop replication on db1095, db1102, db1125 to change triggers - T192926 [production]
06:30 <marostegui@deploy1001> Synchronized wmf-config/db-eqiad.php: Repool db1091 after alter table (duration: 00m 51s) [production]
05:09 <marostegui> Deploy schema change on db1091 T191316 T192926 T89737 T195193 [production]
05:09 <marostegui@deploy1001> Synchronized wmf-config/db-eqiad.php: Depool db1091 for alter table (duration: 00m 52s) [production]
02:45 <l10nupdate@deploy1001> ResourceLoader cache refresh completed at Tue Jun 12 02:45:53 UTC 2018 (duration 10m 18s) [production]