9951-10000 of 10000 results (52ms)
2015-06-08 §
23:11 <ebernhardson> Synchronized php-1.26wmf8/extensions/UploadWizard/: Bump UploadWizard in 1.26wmf8 for evening SWAT (duration: 01m 09s) [production]
22:21 <bblack> depooled cp3030, cp1065 in pybal for ipsec [production]
20:17 <subbu> deployed parsoid sha 131554ba [production]
19:18 <jynus> RAID degradation (disk failure) on s5 master (db1058), no production impact, replacement on the way [production]
17:13 <ottomata> restarted eventlogging services on eventlog1001 after disabling kafka pieces [production]
16:13 <_joe_> powercycling tmh1001, console blank, unresponsive to pings [production]
16:00 <thcipriani> Synchronized commonsuploads.dblist: SWAT: Revert Temporarily re-enable uploads on Marathi Wikipedia, for real [[gerrit:216719]] (duration: 01m 07s) [production]
15:58 <thcipriani> Synchronized commonsuploads.dblist: SWAT: Revert Temporarily re-enable uploads on Marathi Wikipedia [[gerrit:216719]] (duration: 01m 08s) [production]
15:40 <thcipriani> Synchronized php-1.26wmf8/extensions/Cite: SWAT: Revert Do all of Cite's real work during unstrip and followup [[gerrit:216715]] (duration: 01m 08s) [production]
15:19 <Coren> T96063: process halted for now as store/backup is unmovable and on slice5 [production]
15:17 <thcipriani> Synchronized w/static/images/project-logos/pflwiki.png: SWAT: Fix transparency of pflwiki logo [[gerrit:216595]] (duration: 01m 08s) [production]
15:15 <akosiaris> disabled ircecho on neon for a while [production]
14:53 <Coren> T96063: starting pvmove from slice5 to slice2 [production]
14:48 <Coren> T96063: dropped volume slice1 from vg store [production]
14:46 <Coren> T96063: dropped store/project [production]
14:44 <Coren> starting https://phabricator.wikimedia.org/T96063 on labstore1001 [production]
14:24 <jynus> Synchronized wmf-config/db-eqiad.php: depool es1005 (duration: 01m 08s) [production]
14:23 <Coren> rsync in progress between labstore1001:store/backup and labstore1002:backup/backup (at ionice idle) [production]
14:13 <Coren> created store/backup snapshot on labstore1001 for backup copy [production]
13:03 <moritzm> added strongswan_5.3.0-1+wmf2 to jessie-wikimedia on carbon [production]
11:42 <_joe_> purging squid cache on carbon [production]
11:26 <moritzm> updated mc2* to 2:2.8.17-1+deb8u1 [production]
10:55 <jynus> Synchronized wmf-config/db-eqiad.php: repool es1007 (duration: 01m 08s) [production]
10:27 <akosiaris> disabled puppet on uranium, investigating ganglia problems [production]
10:05 <akosiaris> ganglia gmetad problems [production]
05:25 <LocalisationUpdate> ResourceLoader cache refresh completed at Mon Jun 8 05:24:08 UTC 2015 (duration 24m 7s) [production]
02:26 <LocalisationUpdate> completed (1.26wmf8) at 2015-06-08 02:25:12+00:00 [production]
02:21 <l10nupdate> Synchronized php-1.26wmf8/cache/l10n: (no message) (duration: 07m 07s) [production]
2015-06-07 §
23:27 <godog> reboot ms-be2008 sdg failed, xfs unhappy [production]
07:03 <springle> Synchronized wmf-config/db-eqiad.php: repool db1073, warm up (duration: 01m 09s) [production]
05:16 <andrewbogott> we did a whole lot of things to labstore1001 while morebots was away [production]
05:14 <andrewbogott> service nfs-kernel-server restart on labstore1001 [production]
02:26 <LocalisationUpdate> completed (1.26wmf8) at 2015-06-07 02:25:13+00:00 [production]
02:21 <l10nupdate> Synchronized php-1.26wmf8/cache/l10n: (no message) (duration: 07m 09s) [production]
2015-06-06 §
23:46 <subbu> deployed parsoid 5172a446 (cherry-pick of 719c736f) -- hotfix for T101599 [production]
05:48 <LocalisationUpdate> ResourceLoader cache refresh completed at Sat Jun 6 05:47:40 UTC 2015 (duration 47m 39s) [production]
02:31 <LocalisationUpdate> completed (1.26wmf8) at 2015-06-06 02:30:24+00:00 [production]
02:26 <l10nupdate> Synchronized php-1.26wmf8/cache/l10n: (no message) (duration: 07m 10s) [production]
2015-06-05 §
22:42 <godog> powercycle graphite2001, no console no ssh [production]
22:06 <andrewbogott> restarted apache on virt1000 [production]
20:49 <ori> Upgrading hhvm-fss on application servers to 1.1.7; expect brief 5xx spike. [production]
20:14 <demon> Synchronized php-1.26wmf8: live hack (duration: 02m 32s) [production]
20:10 <mutante> apt-get upgrade on terbium [production]
19:52 <godog> bounce redis on rdb1001/rdb1003 to pick up new slave limits [production]
19:51 <mutante> chown root:root / on terbium [production]
19:50 <godog> bounce redis on rdb1002/rdb1004 to pick up new slave limits [production]
19:29 <godog> bounce redis again on rdb1003 after increasing the slave limits more [production]
19:17 <godog> bounce redis on rdb1003 after bumping slave limits [production]
19:07 <godog> redis master logs shows periodic 'cmd=sync scheduled to be closed ASAP for overcoming of output buffer limits.' indicating the slave fails to sync [production]
18:40 <godog> spike in redis network starting at ~15.00 UTC, correlates with ocg failures [production]